var/home/core/zuul-output/0000755000175000017500000000000015111000174014513 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111010773015465 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004434630315111010765017701 0ustar rootrootNov 24 06:54:58 crc systemd[1]: Starting Kubernetes Kubelet... Nov 24 06:54:58 crc restorecon[4676]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 24 06:54:58 crc restorecon[4676]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:54:58 crc restorecon[4676]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:54:58 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 06:54:58 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 06:54:58 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 06:54:58 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 06:54:58 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 06:54:58 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 06:54:58 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 06:54:58 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 06:54:58 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 06:54:58 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 06:54:58 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 06:54:58 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 06:54:58 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 06:54:58 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 06:54:58 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 06:54:58 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:54:58 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:54:58 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:54:58 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:54:58 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:54:59 crc restorecon[4676]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:54:59 crc restorecon[4676]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 24 06:55:00 crc kubenswrapper[4809]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 06:55:00 crc kubenswrapper[4809]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 24 06:55:00 crc kubenswrapper[4809]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 06:55:00 crc kubenswrapper[4809]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 06:55:00 crc kubenswrapper[4809]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 24 06:55:00 crc kubenswrapper[4809]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.599179 4809 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605065 4809 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605097 4809 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605108 4809 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605118 4809 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605128 4809 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605136 4809 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605145 4809 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605154 4809 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605169 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605181 4809 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605199 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605219 4809 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605233 4809 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605243 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605254 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605263 4809 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605272 4809 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605280 4809 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605289 4809 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605297 4809 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605305 4809 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605313 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605321 4809 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605330 4809 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605338 4809 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605346 4809 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605355 4809 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605363 4809 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605371 4809 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605380 4809 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605388 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605400 4809 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605411 4809 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605424 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605434 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605447 4809 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605459 4809 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605469 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605478 4809 feature_gate.go:330] unrecognized feature gate: Example Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605488 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605497 4809 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605506 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605517 4809 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605526 4809 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605538 4809 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605547 4809 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605555 4809 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605565 4809 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605573 4809 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605583 4809 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605591 4809 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605599 4809 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605607 4809 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605615 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605624 4809 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605637 4809 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605647 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605656 4809 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605665 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605673 4809 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605681 4809 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605690 4809 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605697 4809 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605706 4809 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605714 4809 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605723 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605731 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605739 4809 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605747 4809 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605758 4809 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.605766 4809 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.605927 4809 flags.go:64] FLAG: --address="0.0.0.0" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.605944 4809 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.605960 4809 flags.go:64] FLAG: --anonymous-auth="true" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606004 4809 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606017 4809 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606027 4809 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606039 4809 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606051 4809 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606064 4809 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606074 4809 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606085 4809 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606095 4809 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606105 4809 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606115 4809 flags.go:64] FLAG: --cgroup-root="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606125 4809 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606134 4809 flags.go:64] FLAG: --client-ca-file="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606144 4809 flags.go:64] FLAG: --cloud-config="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606153 4809 flags.go:64] FLAG: --cloud-provider="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606163 4809 flags.go:64] FLAG: --cluster-dns="[]" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606176 4809 flags.go:64] FLAG: --cluster-domain="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606188 4809 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606202 4809 flags.go:64] FLAG: --config-dir="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606214 4809 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606227 4809 flags.go:64] FLAG: --container-log-max-files="5" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606242 4809 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606253 4809 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606263 4809 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606273 4809 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606282 4809 flags.go:64] FLAG: --contention-profiling="false" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606292 4809 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606302 4809 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606312 4809 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606322 4809 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606334 4809 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606345 4809 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606355 4809 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606365 4809 flags.go:64] FLAG: --enable-load-reader="false" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606374 4809 flags.go:64] FLAG: --enable-server="true" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606384 4809 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606396 4809 flags.go:64] FLAG: --event-burst="100" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606407 4809 flags.go:64] FLAG: --event-qps="50" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606417 4809 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606427 4809 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606437 4809 flags.go:64] FLAG: --eviction-hard="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606448 4809 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606457 4809 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606468 4809 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606478 4809 flags.go:64] FLAG: --eviction-soft="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606487 4809 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606497 4809 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606507 4809 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606517 4809 flags.go:64] FLAG: --experimental-mounter-path="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606526 4809 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606535 4809 flags.go:64] FLAG: --fail-swap-on="true" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606545 4809 flags.go:64] FLAG: --feature-gates="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606557 4809 flags.go:64] FLAG: --file-check-frequency="20s" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606567 4809 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606577 4809 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606586 4809 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606596 4809 flags.go:64] FLAG: --healthz-port="10248" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606606 4809 flags.go:64] FLAG: --help="false" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606616 4809 flags.go:64] FLAG: --hostname-override="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606625 4809 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606635 4809 flags.go:64] FLAG: --http-check-frequency="20s" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606645 4809 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606654 4809 flags.go:64] FLAG: --image-credential-provider-config="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606663 4809 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606673 4809 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606683 4809 flags.go:64] FLAG: --image-service-endpoint="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606692 4809 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606705 4809 flags.go:64] FLAG: --kube-api-burst="100" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606715 4809 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606726 4809 flags.go:64] FLAG: --kube-api-qps="50" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606736 4809 flags.go:64] FLAG: --kube-reserved="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606782 4809 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606792 4809 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606802 4809 flags.go:64] FLAG: --kubelet-cgroups="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606812 4809 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606821 4809 flags.go:64] FLAG: --lock-file="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606831 4809 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606841 4809 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606851 4809 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606873 4809 flags.go:64] FLAG: --log-json-split-stream="false" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606882 4809 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606892 4809 flags.go:64] FLAG: --log-text-split-stream="false" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606902 4809 flags.go:64] FLAG: --logging-format="text" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606911 4809 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606922 4809 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606932 4809 flags.go:64] FLAG: --manifest-url="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606943 4809 flags.go:64] FLAG: --manifest-url-header="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606957 4809 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.606994 4809 flags.go:64] FLAG: --max-open-files="1000000" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607006 4809 flags.go:64] FLAG: --max-pods="110" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607017 4809 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607027 4809 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607037 4809 flags.go:64] FLAG: --memory-manager-policy="None" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607046 4809 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607056 4809 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607066 4809 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607076 4809 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607097 4809 flags.go:64] FLAG: --node-status-max-images="50" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607107 4809 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607117 4809 flags.go:64] FLAG: --oom-score-adj="-999" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607127 4809 flags.go:64] FLAG: --pod-cidr="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607137 4809 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607152 4809 flags.go:64] FLAG: --pod-manifest-path="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607163 4809 flags.go:64] FLAG: --pod-max-pids="-1" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607177 4809 flags.go:64] FLAG: --pods-per-core="0" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607190 4809 flags.go:64] FLAG: --port="10250" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607204 4809 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607216 4809 flags.go:64] FLAG: --provider-id="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607228 4809 flags.go:64] FLAG: --qos-reserved="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607240 4809 flags.go:64] FLAG: --read-only-port="10255" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607251 4809 flags.go:64] FLAG: --register-node="true" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607261 4809 flags.go:64] FLAG: --register-schedulable="true" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607271 4809 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607287 4809 flags.go:64] FLAG: --registry-burst="10" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607296 4809 flags.go:64] FLAG: --registry-qps="5" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607306 4809 flags.go:64] FLAG: --reserved-cpus="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607317 4809 flags.go:64] FLAG: --reserved-memory="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607329 4809 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607339 4809 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607348 4809 flags.go:64] FLAG: --rotate-certificates="false" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607358 4809 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607368 4809 flags.go:64] FLAG: --runonce="false" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607377 4809 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607387 4809 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607397 4809 flags.go:64] FLAG: --seccomp-default="false" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607407 4809 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607417 4809 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607427 4809 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607437 4809 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607447 4809 flags.go:64] FLAG: --storage-driver-password="root" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607456 4809 flags.go:64] FLAG: --storage-driver-secure="false" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607466 4809 flags.go:64] FLAG: --storage-driver-table="stats" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607475 4809 flags.go:64] FLAG: --storage-driver-user="root" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607485 4809 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607495 4809 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607504 4809 flags.go:64] FLAG: --system-cgroups="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607514 4809 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607529 4809 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607539 4809 flags.go:64] FLAG: --tls-cert-file="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607549 4809 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607561 4809 flags.go:64] FLAG: --tls-min-version="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607572 4809 flags.go:64] FLAG: --tls-private-key-file="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607581 4809 flags.go:64] FLAG: --topology-manager-policy="none" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607591 4809 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607601 4809 flags.go:64] FLAG: --topology-manager-scope="container" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607610 4809 flags.go:64] FLAG: --v="2" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607622 4809 flags.go:64] FLAG: --version="false" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607634 4809 flags.go:64] FLAG: --vmodule="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607645 4809 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.607656 4809 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.607869 4809 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.607880 4809 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.607891 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.607900 4809 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.607909 4809 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.607919 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.607936 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.607960 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608009 4809 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608020 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608031 4809 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608042 4809 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608055 4809 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608065 4809 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608075 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608085 4809 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608096 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608106 4809 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608117 4809 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608131 4809 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608144 4809 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608156 4809 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608167 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608179 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608204 4809 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608229 4809 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608242 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608253 4809 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608264 4809 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608273 4809 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608282 4809 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608290 4809 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608299 4809 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608312 4809 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608321 4809 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608331 4809 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608339 4809 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608349 4809 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608358 4809 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608367 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608377 4809 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608385 4809 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608395 4809 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608404 4809 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608413 4809 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608423 4809 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608435 4809 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608444 4809 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608454 4809 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608463 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608472 4809 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608481 4809 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608490 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608498 4809 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608506 4809 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608515 4809 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608523 4809 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608531 4809 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608540 4809 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608548 4809 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608556 4809 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608566 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608574 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608582 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608590 4809 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608599 4809 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608607 4809 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608615 4809 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608624 4809 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608632 4809 feature_gate.go:330] unrecognized feature gate: Example Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.608641 4809 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.610201 4809 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.622741 4809 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.622838 4809 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.623781 4809 feature_gate.go:330] unrecognized feature gate: Example Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624222 4809 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624239 4809 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624246 4809 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624252 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624259 4809 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624266 4809 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624272 4809 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624278 4809 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624283 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624288 4809 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624294 4809 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624299 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624304 4809 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624309 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624315 4809 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624320 4809 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624325 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624331 4809 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624336 4809 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624341 4809 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624346 4809 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624356 4809 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624361 4809 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624366 4809 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624372 4809 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624377 4809 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624381 4809 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624388 4809 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624395 4809 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624401 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624408 4809 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624414 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624421 4809 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624427 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624432 4809 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624437 4809 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624442 4809 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624448 4809 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624453 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624458 4809 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624463 4809 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624468 4809 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624473 4809 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624478 4809 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624483 4809 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624488 4809 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624493 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624498 4809 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624520 4809 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624525 4809 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624532 4809 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624537 4809 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624542 4809 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624547 4809 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624552 4809 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624557 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624563 4809 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624568 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624574 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624579 4809 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624584 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624589 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624594 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624599 4809 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624605 4809 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624611 4809 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624618 4809 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624624 4809 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624631 4809 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624636 4809 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.624645 4809 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624791 4809 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624800 4809 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624805 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624811 4809 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624815 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624820 4809 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624825 4809 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624830 4809 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624835 4809 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624840 4809 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624845 4809 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624849 4809 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624855 4809 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624860 4809 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624866 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624872 4809 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624880 4809 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624886 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624892 4809 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624899 4809 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624904 4809 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624909 4809 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624914 4809 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624919 4809 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624924 4809 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624930 4809 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624935 4809 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624940 4809 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624944 4809 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624950 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624954 4809 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624959 4809 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624981 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624986 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624991 4809 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.624996 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625001 4809 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625007 4809 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625013 4809 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625018 4809 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625023 4809 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625028 4809 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625034 4809 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625041 4809 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625046 4809 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625052 4809 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625057 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625071 4809 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625077 4809 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625082 4809 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625087 4809 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625092 4809 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625097 4809 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625104 4809 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625110 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625116 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625122 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625128 4809 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625133 4809 feature_gate.go:330] unrecognized feature gate: Example Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625139 4809 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625143 4809 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625148 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625153 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625158 4809 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625163 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625168 4809 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625173 4809 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625178 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625182 4809 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625187 4809 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.625192 4809 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.625200 4809 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.625373 4809 server.go:940] "Client rotation is on, will bootstrap in background" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.630656 4809 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.630929 4809 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.633162 4809 server.go:997] "Starting client certificate rotation" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.633213 4809 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.633451 4809 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-15 17:30:52.542021175 +0000 UTC Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.633594 4809 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.661270 4809 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.664947 4809 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 06:55:00 crc kubenswrapper[4809]: E1124 06:55:00.669439 4809 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.159:6443: connect: connection refused" logger="UnhandledError" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.689297 4809 log.go:25] "Validated CRI v1 runtime API" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.735998 4809 log.go:25] "Validated CRI v1 image API" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.739128 4809 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.748704 4809 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-24-06-49-44-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.748771 4809 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.785113 4809 manager.go:217] Machine: {Timestamp:2025-11-24 06:55:00.779960596 +0000 UTC m=+0.680552321 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:c636b03a-5cf2-4b68-8edd-3198ca3d19e0 BootID:2f7265f8-062f-41c9-817e-b7f54d138dae Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:ea:6f:0a Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:ea:6f:0a Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:a9:7f:d0 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:8c:29:30 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:c5:35:94 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:48:dc:9c Speed:-1 Mtu:1496} {Name:eth10 MacAddress:86:fb:e8:86:61:12 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:d2:14:c0:a7:51:1c Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.785422 4809 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.785610 4809 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.786950 4809 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.787191 4809 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.787252 4809 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.787478 4809 topology_manager.go:138] "Creating topology manager with none policy" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.787493 4809 container_manager_linux.go:303] "Creating device plugin manager" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.788433 4809 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.788476 4809 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.788675 4809 state_mem.go:36] "Initialized new in-memory state store" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.788768 4809 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.795494 4809 kubelet.go:418] "Attempting to sync node with API server" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.795523 4809 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.795550 4809 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.795565 4809 kubelet.go:324] "Adding apiserver pod source" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.795582 4809 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.801146 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Nov 24 06:55:00 crc kubenswrapper[4809]: E1124 06:55:00.801249 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.159:6443: connect: connection refused" logger="UnhandledError" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.801438 4809 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.802417 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Nov 24 06:55:00 crc kubenswrapper[4809]: E1124 06:55:00.802566 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.159:6443: connect: connection refused" logger="UnhandledError" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.802716 4809 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.806911 4809 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.808925 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.809005 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.809023 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.809038 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.809069 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.809084 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.809105 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.809130 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.809148 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.809163 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.809189 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.809203 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.811123 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.811899 4809 server.go:1280] "Started kubelet" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.813265 4809 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.813262 4809 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 24 06:55:00 crc systemd[1]: Started Kubernetes Kubelet. Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.814273 4809 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.814305 4809 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.814623 4809 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 07:11:47.136408011 +0000 UTC Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.814658 4809 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 576h16m46.32175253s for next certificate rotation Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.814649 4809 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.814688 4809 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 24 06:55:00 crc kubenswrapper[4809]: E1124 06:55:00.814735 4809 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.815067 4809 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.816021 4809 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.816084 4809 factory.go:55] Registering systemd factory Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.816106 4809 factory.go:221] Registration of the systemd container factory successfully Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.816206 4809 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.817539 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Nov 24 06:55:00 crc kubenswrapper[4809]: E1124 06:55:00.817632 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.159:6443: connect: connection refused" logger="UnhandledError" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.817723 4809 factory.go:153] Registering CRI-O factory Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.817759 4809 factory.go:221] Registration of the crio container factory successfully Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.817874 4809 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.817914 4809 factory.go:103] Registering Raw factory Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.817944 4809 manager.go:1196] Started watching for new ooms in manager Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.818295 4809 server.go:460] "Adding debug handlers to kubelet server" Nov 24 06:55:00 crc kubenswrapper[4809]: E1124 06:55:00.825329 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.159:6443: connect: connection refused" interval="200ms" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.825895 4809 manager.go:319] Starting recovery of all containers Nov 24 06:55:00 crc kubenswrapper[4809]: E1124 06:55:00.831382 4809 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.159:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187adeea084b93ff default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-24 06:55:00.811854847 +0000 UTC m=+0.712446592,LastTimestamp:2025-11-24 06:55:00.811854847 +0000 UTC m=+0.712446592,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841158 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841219 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841237 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841251 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841274 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841300 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841314 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841327 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841343 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841357 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841369 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841383 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841395 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841411 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841423 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841438 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841453 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841465 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841477 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841491 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841505 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841519 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841531 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841576 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841592 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841610 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841625 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841638 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841653 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841664 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841701 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841714 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841725 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841735 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841746 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841757 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841772 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841784 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841796 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841808 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841819 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841833 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841849 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841864 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841882 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841898 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841915 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841928 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841941 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.841953 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842016 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842033 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842049 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842062 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842076 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842087 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842099 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842111 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842123 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842137 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842148 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842160 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842176 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842219 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842240 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842253 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842264 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842277 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842290 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842302 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842315 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842328 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842340 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842352 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842364 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842375 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842387 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842401 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842429 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842455 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842473 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842488 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842503 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842519 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842539 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842556 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842573 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842590 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842604 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842619 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842634 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842648 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842664 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842679 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842695 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842712 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842728 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842743 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842759 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842775 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842789 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842804 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842820 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842834 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842858 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842875 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842895 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842945 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.842988 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843011 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843034 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843054 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843072 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843088 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843105 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843121 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843137 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843153 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843172 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843190 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843208 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843224 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843295 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843317 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843334 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843350 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843366 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843384 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843428 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843443 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843459 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843475 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843492 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843509 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843525 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843542 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843558 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.843575 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846338 4809 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846417 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846443 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846466 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846478 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846493 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846508 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846521 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846534 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846548 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846561 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846578 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846595 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846607 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846622 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846638 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846653 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846668 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846680 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846695 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846710 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846723 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846736 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846747 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846761 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846775 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846788 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846800 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846815 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846827 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846841 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846853 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846865 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846876 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846887 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846899 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846933 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846934 4809 manager.go:324] Recovery completed Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.846949 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847001 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847014 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847031 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847051 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847070 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847087 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847102 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847116 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847130 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847145 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847163 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847178 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847193 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847209 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847223 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847239 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847254 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847293 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847311 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847327 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847343 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847357 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847371 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847386 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847400 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847414 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847430 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847445 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847461 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847479 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847493 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847508 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847520 4809 reconstruct.go:97] "Volume reconstruction finished" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.847531 4809 reconciler.go:26] "Reconciler: start to sync state" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.858092 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.859572 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.859607 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.859619 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.860511 4809 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.860533 4809 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.860557 4809 state_mem.go:36] "Initialized new in-memory state store" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.888088 4809 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.889849 4809 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.889903 4809 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.889931 4809 kubelet.go:2335] "Starting kubelet main sync loop" Nov 24 06:55:00 crc kubenswrapper[4809]: E1124 06:55:00.889992 4809 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.890550 4809 policy_none.go:49] "None policy: Start" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.891170 4809 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.891203 4809 state_mem.go:35] "Initializing new in-memory state store" Nov 24 06:55:00 crc kubenswrapper[4809]: W1124 06:55:00.892592 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Nov 24 06:55:00 crc kubenswrapper[4809]: E1124 06:55:00.892765 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.159:6443: connect: connection refused" logger="UnhandledError" Nov 24 06:55:00 crc kubenswrapper[4809]: E1124 06:55:00.915871 4809 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.962340 4809 manager.go:334] "Starting Device Plugin manager" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.962413 4809 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.962432 4809 server.go:79] "Starting device plugin registration server" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.963025 4809 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.963051 4809 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.964536 4809 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.964643 4809 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.964650 4809 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 24 06:55:00 crc kubenswrapper[4809]: E1124 06:55:00.970636 4809 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.990293 4809 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.990571 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.992063 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.992097 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.992108 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.992208 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.992456 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.992499 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.992868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.992898 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.992909 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.993010 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.993091 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.993130 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.993397 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.993425 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.993435 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.993733 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.993779 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.993789 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.993814 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.993831 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.993876 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.993896 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.994047 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.994121 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.994596 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.994619 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.994655 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.994748 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.994764 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.994769 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.994824 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.994847 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.994773 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.995514 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.995537 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.995549 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.995591 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.995608 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.995618 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.995764 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.995789 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.996346 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.996370 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:00 crc kubenswrapper[4809]: I1124 06:55:00.996380 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:01 crc kubenswrapper[4809]: E1124 06:55:01.026635 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.159:6443: connect: connection refused" interval="400ms" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.049497 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.049523 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.049540 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.049555 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.049571 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.049596 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.049643 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.049671 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.049691 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.049714 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.049766 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.049811 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.049841 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.049871 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.049900 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.063544 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.064521 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.064548 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.064558 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.064579 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 06:55:01 crc kubenswrapper[4809]: E1124 06:55:01.064906 4809 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.159:6443: connect: connection refused" node="crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.150639 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.150718 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.150766 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.150787 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.150807 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.150847 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.150883 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.150888 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.150927 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.150886 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.150857 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.150848 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.151035 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.151066 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.151077 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.151092 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.151152 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.151159 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.151184 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.151188 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.151207 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.151223 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.151229 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.151254 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.151252 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.151280 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.151288 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.151321 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.151316 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.151291 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.265608 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.266688 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.266744 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.266760 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.266791 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 06:55:01 crc kubenswrapper[4809]: E1124 06:55:01.267323 4809 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.159:6443: connect: connection refused" node="crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.325000 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.339539 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.357486 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.365106 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.370658 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:55:01 crc kubenswrapper[4809]: W1124 06:55:01.377149 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-40fe76ace9dffd64843748bba91b2c907be0290dc76e5f29a9df7603d28f0100 WatchSource:0}: Error finding container 40fe76ace9dffd64843748bba91b2c907be0290dc76e5f29a9df7603d28f0100: Status 404 returned error can't find the container with id 40fe76ace9dffd64843748bba91b2c907be0290dc76e5f29a9df7603d28f0100 Nov 24 06:55:01 crc kubenswrapper[4809]: W1124 06:55:01.378474 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-6e246dbaf7e84313fff0aea36577389897e1d218260e3f574175d70a783f7164 WatchSource:0}: Error finding container 6e246dbaf7e84313fff0aea36577389897e1d218260e3f574175d70a783f7164: Status 404 returned error can't find the container with id 6e246dbaf7e84313fff0aea36577389897e1d218260e3f574175d70a783f7164 Nov 24 06:55:01 crc kubenswrapper[4809]: W1124 06:55:01.388324 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-3e19522bc6787f2164807ef71f26ce4dda894b8eeafa0aed82194af549ff8375 WatchSource:0}: Error finding container 3e19522bc6787f2164807ef71f26ce4dda894b8eeafa0aed82194af549ff8375: Status 404 returned error can't find the container with id 3e19522bc6787f2164807ef71f26ce4dda894b8eeafa0aed82194af549ff8375 Nov 24 06:55:01 crc kubenswrapper[4809]: W1124 06:55:01.394034 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-ebb5ab7cd49c8903a93bca26a183de875b22bbd72fd033e738c733bf106f7118 WatchSource:0}: Error finding container ebb5ab7cd49c8903a93bca26a183de875b22bbd72fd033e738c733bf106f7118: Status 404 returned error can't find the container with id ebb5ab7cd49c8903a93bca26a183de875b22bbd72fd033e738c733bf106f7118 Nov 24 06:55:01 crc kubenswrapper[4809]: W1124 06:55:01.396355 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-7c15135ecc297d179bb96b0e4d55cee50f734b440d1f2e857cc559670b6dbc21 WatchSource:0}: Error finding container 7c15135ecc297d179bb96b0e4d55cee50f734b440d1f2e857cc559670b6dbc21: Status 404 returned error can't find the container with id 7c15135ecc297d179bb96b0e4d55cee50f734b440d1f2e857cc559670b6dbc21 Nov 24 06:55:01 crc kubenswrapper[4809]: E1124 06:55:01.428610 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.159:6443: connect: connection refused" interval="800ms" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.667895 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.669542 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.669587 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.669603 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.669632 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 06:55:01 crc kubenswrapper[4809]: E1124 06:55:01.670150 4809 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.159:6443: connect: connection refused" node="crc" Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.816959 4809 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.894550 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"40fe76ace9dffd64843748bba91b2c907be0290dc76e5f29a9df7603d28f0100"} Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.896234 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6e246dbaf7e84313fff0aea36577389897e1d218260e3f574175d70a783f7164"} Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.897155 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7c15135ecc297d179bb96b0e4d55cee50f734b440d1f2e857cc559670b6dbc21"} Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.898073 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ebb5ab7cd49c8903a93bca26a183de875b22bbd72fd033e738c733bf106f7118"} Nov 24 06:55:01 crc kubenswrapper[4809]: I1124 06:55:01.898978 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3e19522bc6787f2164807ef71f26ce4dda894b8eeafa0aed82194af549ff8375"} Nov 24 06:55:01 crc kubenswrapper[4809]: W1124 06:55:01.969835 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Nov 24 06:55:01 crc kubenswrapper[4809]: E1124 06:55:01.969897 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.159:6443: connect: connection refused" logger="UnhandledError" Nov 24 06:55:02 crc kubenswrapper[4809]: W1124 06:55:02.038848 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Nov 24 06:55:02 crc kubenswrapper[4809]: E1124 06:55:02.038925 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.159:6443: connect: connection refused" logger="UnhandledError" Nov 24 06:55:02 crc kubenswrapper[4809]: W1124 06:55:02.087590 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Nov 24 06:55:02 crc kubenswrapper[4809]: E1124 06:55:02.087676 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.159:6443: connect: connection refused" logger="UnhandledError" Nov 24 06:55:02 crc kubenswrapper[4809]: W1124 06:55:02.212779 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Nov 24 06:55:02 crc kubenswrapper[4809]: E1124 06:55:02.212894 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.159:6443: connect: connection refused" logger="UnhandledError" Nov 24 06:55:02 crc kubenswrapper[4809]: E1124 06:55:02.230393 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.159:6443: connect: connection refused" interval="1.6s" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.470249 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.473250 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.473272 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.473281 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.473302 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 06:55:02 crc kubenswrapper[4809]: E1124 06:55:02.473620 4809 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.159:6443: connect: connection refused" node="crc" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.817906 4809 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.831820 4809 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 24 06:55:02 crc kubenswrapper[4809]: E1124 06:55:02.833240 4809 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.159:6443: connect: connection refused" logger="UnhandledError" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.904245 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6" exitCode=0 Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.904361 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.904373 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6"} Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.905230 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.905263 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.905275 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.907053 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.908049 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.908075 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.908084 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.908929 4809 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e363c3138adb287372250ef36be1d4f4058a9a72d1b98e5772a76f46854e122e" exitCode=0 Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.909007 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e363c3138adb287372250ef36be1d4f4058a9a72d1b98e5772a76f46854e122e"} Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.909121 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.909823 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.909840 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.909854 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.910814 4809 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="fb1a74c3f4524f23702cf96f82ca151699897f7a1fe9eafdb035962e93ed897d" exitCode=0 Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.910842 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"fb1a74c3f4524f23702cf96f82ca151699897f7a1fe9eafdb035962e93ed897d"} Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.910900 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.911678 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.911700 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.911709 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.912633 4809 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="66bb948c15a1089cc29548df9da9f67b1f554d907a02f6076a0da38cabca9670" exitCode=0 Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.912760 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"66bb948c15a1089cc29548df9da9f67b1f554d907a02f6076a0da38cabca9670"} Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.912775 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.913492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.913521 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.913531 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.921381 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175"} Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.921413 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4"} Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.921425 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60"} Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.921438 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33"} Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.921655 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.923033 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.923082 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:02 crc kubenswrapper[4809]: I1124 06:55:02.923092 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.816593 4809 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Nov 24 06:55:03 crc kubenswrapper[4809]: E1124 06:55:03.831165 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.159:6443: connect: connection refused" interval="3.2s" Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.926719 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2"} Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.926779 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50"} Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.926796 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11"} Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.926809 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3"} Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.926816 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.926821 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2"} Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.927639 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.927669 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.927679 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.928321 4809 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="6c173093d59b165e37daa3875a107ae11273b7f67a08eb03791888cd3b40dcb8" exitCode=0 Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.928388 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"6c173093d59b165e37daa3875a107ae11273b7f67a08eb03791888cd3b40dcb8"} Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.928428 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.929203 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.929256 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.929275 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.929780 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"6e21b2b68cf22971d6cedd4d455d80409d8d00e569d5b2ce0319b15e79c91d0f"} Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.929789 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.930419 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.930434 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.930443 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.931809 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e70c7310e5734c08f62e5732d069aed4eb9ccd5cc0e8ae546950871772f6a27d"} Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.931862 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"5376dc11940017714443d1a4cf6c61f323eb35b41d1ce9b76d73137465f18c83"} Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.931887 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"80b7097545fca095740245eb7373a8ad9325bfde309725cf0b0db792bbd44df7"} Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.931827 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.931837 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.932905 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.932939 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.932950 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.932952 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.933017 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:03 crc kubenswrapper[4809]: I1124 06:55:03.933093 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.073753 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.074677 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.074709 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.074721 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.074748 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 06:55:04 crc kubenswrapper[4809]: E1124 06:55:04.075193 4809 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.159:6443: connect: connection refused" node="crc" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.615879 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.935810 4809 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9377df003b08b18df163f2c6165e350a6c13408dbe9af8395754ec8b1d2b26f1" exitCode=0 Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.935902 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.935989 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.936010 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9377df003b08b18df163f2c6165e350a6c13408dbe9af8395754ec8b1d2b26f1"} Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.936066 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.936035 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.936118 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.936159 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.937176 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.937200 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.937212 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.937381 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.937408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.937421 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.937466 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.937490 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.937500 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.937536 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.937561 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:04 crc kubenswrapper[4809]: I1124 06:55:04.937572 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:05 crc kubenswrapper[4809]: I1124 06:55:05.941823 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"baf4ab4375a6db95deefffb75e51467873f4d1e04c5fd983ab5ea83e40550b89"} Nov 24 06:55:05 crc kubenswrapper[4809]: I1124 06:55:05.941889 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:05 crc kubenswrapper[4809]: I1124 06:55:05.941897 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"04649fa3e34fcd177289ea8d5c2b8db4e980f76eca41d2296d6beb6a42ce5d6f"} Nov 24 06:55:05 crc kubenswrapper[4809]: I1124 06:55:05.941914 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"cdb84a4ee15521f70a3580b9cc45d48a82efd1c29097a7cc6957a58f02393a48"} Nov 24 06:55:05 crc kubenswrapper[4809]: I1124 06:55:05.941926 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3792f62811d06ba36e30cc563657d8f0cf5d835d8617a16d6d3fc9c870a0777b"} Nov 24 06:55:05 crc kubenswrapper[4809]: I1124 06:55:05.941938 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"22f6e1a25b785cf38c517ee7636b71d6c91d95767d15ed02bef2c03c98e9c715"} Nov 24 06:55:05 crc kubenswrapper[4809]: I1124 06:55:05.941900 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:05 crc kubenswrapper[4809]: I1124 06:55:05.943415 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:05 crc kubenswrapper[4809]: I1124 06:55:05.943446 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:05 crc kubenswrapper[4809]: I1124 06:55:05.943418 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:05 crc kubenswrapper[4809]: I1124 06:55:05.943470 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:05 crc kubenswrapper[4809]: I1124 06:55:05.943481 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:05 crc kubenswrapper[4809]: I1124 06:55:05.943455 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:06 crc kubenswrapper[4809]: I1124 06:55:06.918120 4809 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 24 06:55:06 crc kubenswrapper[4809]: I1124 06:55:06.944173 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:06 crc kubenswrapper[4809]: I1124 06:55:06.944855 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:06 crc kubenswrapper[4809]: I1124 06:55:06.944883 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:06 crc kubenswrapper[4809]: I1124 06:55:06.944891 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:07 crc kubenswrapper[4809]: I1124 06:55:07.195527 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:55:07 crc kubenswrapper[4809]: I1124 06:55:07.195735 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:07 crc kubenswrapper[4809]: I1124 06:55:07.196817 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:07 crc kubenswrapper[4809]: I1124 06:55:07.196847 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:07 crc kubenswrapper[4809]: I1124 06:55:07.196856 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:07 crc kubenswrapper[4809]: I1124 06:55:07.275346 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:07 crc kubenswrapper[4809]: I1124 06:55:07.276338 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:07 crc kubenswrapper[4809]: I1124 06:55:07.276379 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:07 crc kubenswrapper[4809]: I1124 06:55:07.276387 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:07 crc kubenswrapper[4809]: I1124 06:55:07.276428 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 06:55:07 crc kubenswrapper[4809]: I1124 06:55:07.614937 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 06:55:07 crc kubenswrapper[4809]: I1124 06:55:07.615220 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:07 crc kubenswrapper[4809]: I1124 06:55:07.616900 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:07 crc kubenswrapper[4809]: I1124 06:55:07.617007 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:07 crc kubenswrapper[4809]: I1124 06:55:07.617030 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:07 crc kubenswrapper[4809]: I1124 06:55:07.694168 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 24 06:55:07 crc kubenswrapper[4809]: I1124 06:55:07.946470 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:07 crc kubenswrapper[4809]: I1124 06:55:07.947266 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:07 crc kubenswrapper[4809]: I1124 06:55:07.947309 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:07 crc kubenswrapper[4809]: I1124 06:55:07.947326 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:07 crc kubenswrapper[4809]: I1124 06:55:07.982865 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 24 06:55:08 crc kubenswrapper[4809]: I1124 06:55:08.948960 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:08 crc kubenswrapper[4809]: I1124 06:55:08.949796 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:08 crc kubenswrapper[4809]: I1124 06:55:08.949847 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:08 crc kubenswrapper[4809]: I1124 06:55:08.949860 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:09 crc kubenswrapper[4809]: I1124 06:55:09.742392 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:55:09 crc kubenswrapper[4809]: I1124 06:55:09.742769 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:09 crc kubenswrapper[4809]: I1124 06:55:09.744254 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:09 crc kubenswrapper[4809]: I1124 06:55:09.744315 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:09 crc kubenswrapper[4809]: I1124 06:55:09.744333 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:09 crc kubenswrapper[4809]: I1124 06:55:09.912615 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:55:09 crc kubenswrapper[4809]: I1124 06:55:09.951444 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:09 crc kubenswrapper[4809]: I1124 06:55:09.952508 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:09 crc kubenswrapper[4809]: I1124 06:55:09.952561 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:09 crc kubenswrapper[4809]: I1124 06:55:09.952573 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:10 crc kubenswrapper[4809]: E1124 06:55:10.970731 4809 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 06:55:11 crc kubenswrapper[4809]: I1124 06:55:11.472865 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:55:11 crc kubenswrapper[4809]: I1124 06:55:11.473064 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:11 crc kubenswrapper[4809]: I1124 06:55:11.474378 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:11 crc kubenswrapper[4809]: I1124 06:55:11.474414 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:11 crc kubenswrapper[4809]: I1124 06:55:11.474426 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:11 crc kubenswrapper[4809]: I1124 06:55:11.478337 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:55:11 crc kubenswrapper[4809]: I1124 06:55:11.954542 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:11 crc kubenswrapper[4809]: I1124 06:55:11.955392 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:11 crc kubenswrapper[4809]: I1124 06:55:11.955419 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:11 crc kubenswrapper[4809]: I1124 06:55:11.955429 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:11 crc kubenswrapper[4809]: I1124 06:55:11.961268 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:55:12 crc kubenswrapper[4809]: I1124 06:55:12.742885 4809 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 06:55:12 crc kubenswrapper[4809]: I1124 06:55:12.742976 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 06:55:12 crc kubenswrapper[4809]: I1124 06:55:12.773227 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:55:12 crc kubenswrapper[4809]: I1124 06:55:12.956369 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:12 crc kubenswrapper[4809]: I1124 06:55:12.957134 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:12 crc kubenswrapper[4809]: I1124 06:55:12.957176 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:12 crc kubenswrapper[4809]: I1124 06:55:12.957192 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:13 crc kubenswrapper[4809]: I1124 06:55:13.061339 4809 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 24 06:55:13 crc kubenswrapper[4809]: I1124 06:55:13.061430 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 24 06:55:13 crc kubenswrapper[4809]: I1124 06:55:13.602060 4809 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 24 06:55:13 crc kubenswrapper[4809]: I1124 06:55:13.602159 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 24 06:55:13 crc kubenswrapper[4809]: I1124 06:55:13.959361 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:13 crc kubenswrapper[4809]: I1124 06:55:13.960819 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:13 crc kubenswrapper[4809]: I1124 06:55:13.960868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:13 crc kubenswrapper[4809]: I1124 06:55:13.960879 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:14 crc kubenswrapper[4809]: W1124 06:55:14.619187 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 24 06:55:14 crc kubenswrapper[4809]: I1124 06:55:14.619492 4809 trace.go:236] Trace[66146967]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 06:55:04.618) (total time: 10001ms): Nov 24 06:55:14 crc kubenswrapper[4809]: Trace[66146967]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10000ms (06:55:14.619) Nov 24 06:55:14 crc kubenswrapper[4809]: Trace[66146967]: [10.001245381s] [10.001245381s] END Nov 24 06:55:14 crc kubenswrapper[4809]: E1124 06:55:14.619725 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 24 06:55:14 crc kubenswrapper[4809]: I1124 06:55:14.816821 4809 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 24 06:55:14 crc kubenswrapper[4809]: E1124 06:55:14.900609 4809 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{crc.187adeea084b93ff default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-24 06:55:00.811854847 +0000 UTC m=+0.712446592,LastTimestamp:2025-11-24 06:55:00.811854847 +0000 UTC m=+0.712446592,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 24 06:55:15 crc kubenswrapper[4809]: I1124 06:55:15.060268 4809 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 24 06:55:15 crc kubenswrapper[4809]: I1124 06:55:15.060349 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 24 06:55:15 crc kubenswrapper[4809]: I1124 06:55:15.063906 4809 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 24 06:55:15 crc kubenswrapper[4809]: I1124 06:55:15.064102 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 24 06:55:17 crc kubenswrapper[4809]: I1124 06:55:17.204281 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:55:17 crc kubenswrapper[4809]: I1124 06:55:17.205792 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:17 crc kubenswrapper[4809]: I1124 06:55:17.207299 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:17 crc kubenswrapper[4809]: I1124 06:55:17.207408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:17 crc kubenswrapper[4809]: I1124 06:55:17.207430 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:17 crc kubenswrapper[4809]: I1124 06:55:17.212099 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:55:17 crc kubenswrapper[4809]: I1124 06:55:17.968956 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:17 crc kubenswrapper[4809]: I1124 06:55:17.969789 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:17 crc kubenswrapper[4809]: I1124 06:55:17.969830 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:17 crc kubenswrapper[4809]: I1124 06:55:17.969840 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:18 crc kubenswrapper[4809]: I1124 06:55:18.013055 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 24 06:55:18 crc kubenswrapper[4809]: I1124 06:55:18.013215 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:18 crc kubenswrapper[4809]: I1124 06:55:18.014370 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:18 crc kubenswrapper[4809]: I1124 06:55:18.014431 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:18 crc kubenswrapper[4809]: I1124 06:55:18.014482 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:18 crc kubenswrapper[4809]: I1124 06:55:18.035609 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 24 06:55:18 crc kubenswrapper[4809]: I1124 06:55:18.972250 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:18 crc kubenswrapper[4809]: I1124 06:55:18.973931 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:18 crc kubenswrapper[4809]: I1124 06:55:18.974030 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:18 crc kubenswrapper[4809]: I1124 06:55:18.974043 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.489167 4809 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.806782 4809 apiserver.go:52] "Watching apiserver" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.827387 4809 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.827748 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.828405 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:19 crc kubenswrapper[4809]: E1124 06:55:19.828488 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.828559 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.828835 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:19 crc kubenswrapper[4809]: E1124 06:55:19.828886 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.828933 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:19 crc kubenswrapper[4809]: E1124 06:55:19.828961 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.829044 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.829345 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.833152 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.833826 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.835697 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.835744 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.835938 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.836121 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.836110 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.836630 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.839769 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.868534 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.888673 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.904661 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.915716 4809 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.918541 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.928986 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.938855 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.946697 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.954549 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.962349 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.972864 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.983412 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:19 crc kubenswrapper[4809]: I1124 06:55:19.990516 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.049344 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.050272 4809 trace.go:236] Trace[45257574]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 06:55:05.265) (total time: 14784ms): Nov 24 06:55:20 crc kubenswrapper[4809]: Trace[45257574]: ---"Objects listed" error: 14784ms (06:55:20.050) Nov 24 06:55:20 crc kubenswrapper[4809]: Trace[45257574]: [14.784510475s] [14.784510475s] END Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.050300 4809 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.050448 4809 trace.go:236] Trace[1551994824]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 06:55:05.075) (total time: 14975ms): Nov 24 06:55:20 crc kubenswrapper[4809]: Trace[1551994824]: ---"Objects listed" error: 14975ms (06:55:20.050) Nov 24 06:55:20 crc kubenswrapper[4809]: Trace[1551994824]: [14.975088021s] [14.975088021s] END Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.050483 4809 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.052240 4809 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.052790 4809 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.055194 4809 trace.go:236] Trace[1011898324]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 06:55:05.071) (total time: 14983ms): Nov 24 06:55:20 crc kubenswrapper[4809]: Trace[1011898324]: ---"Objects listed" error: 14983ms (06:55:20.054) Nov 24 06:55:20 crc kubenswrapper[4809]: Trace[1011898324]: [14.983962766s] [14.983962766s] END Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.055241 4809 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.069276 4809 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.097659 4809 csr.go:261] certificate signing request csr-hl9pc is approved, waiting to be issued Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.106805 4809 csr.go:257] certificate signing request csr-hl9pc is issued Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.153744 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.153784 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.153807 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.153827 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.153844 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.153858 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.153872 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.153887 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.153903 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.153927 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.153944 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.153972 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.153988 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154002 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154017 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154032 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154111 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154132 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154149 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154167 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154184 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154199 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154214 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154253 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154291 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154336 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154352 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154367 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154383 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154398 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154413 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154427 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154468 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154449 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154484 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154503 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154461 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154519 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154536 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154581 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154575 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154597 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154613 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154629 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154645 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154665 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154680 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154696 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154713 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154728 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154742 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154758 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154776 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154793 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154810 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154825 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154840 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154854 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154869 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154913 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154931 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154945 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154960 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154989 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155005 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155027 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155042 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155056 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155071 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155086 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155102 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155116 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155134 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155149 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155167 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155182 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155196 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155213 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155230 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155245 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155260 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155274 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155318 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155336 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155350 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155366 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155381 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155396 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155411 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155427 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155441 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155459 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155477 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155492 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155508 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155525 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155540 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154750 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.154996 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155051 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155111 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155109 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155104 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155143 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155254 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155289 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155376 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155397 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155447 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155452 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155477 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155514 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155606 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155654 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155791 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.155812 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.156043 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.156988 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.156057 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.156192 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.156246 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.156272 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.156266 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.156319 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.156359 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.157045 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.158278 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.158318 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.158371 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.158723 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.158752 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.158851 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.158874 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.159734 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.159778 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.159921 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.159951 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.156498 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.156539 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.156658 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.156762 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.156814 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.156984 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.157201 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.157347 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.157555 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.157696 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.157803 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.157803 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.157982 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.157991 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.157961 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.157998 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.158047 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.158139 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.158361 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.158415 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.158507 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.158545 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.158786 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.158797 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.158863 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.160069 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.160713 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.161153 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.161151 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.161195 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.161218 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.161238 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.161521 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.161480 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.161731 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.162631 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.164424 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.165025 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.165063 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.165114 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.165140 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.165161 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.165205 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.165230 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.165274 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.165298 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.165319 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.165362 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.165385 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.165406 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.165457 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.165481 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.165504 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166223 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166264 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166297 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166325 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166357 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166386 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166417 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166444 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166475 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166504 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166530 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166562 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166592 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166622 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166651 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166678 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166705 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166734 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166765 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166797 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.167058 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.167124 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.167157 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.167190 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.167222 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.167276 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.167310 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.167344 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.167378 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.167412 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.167446 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.167479 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.167511 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.167547 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.167582 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.167615 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.165032 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.165349 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166026 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166123 4809 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:43222->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.167743 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:43222->192.168.126.11:17697: read: connection reset by peer" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.167930 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.167958 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.167993 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.170116 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.170545 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.170607 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.170681 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.170823 4809 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.171519 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.167647 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.171708 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.171786 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.171855 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.171927 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.172025 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.172091 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.172167 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.172250 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.172321 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.172389 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.172457 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.172541 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.172605 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.172682 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.172758 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.172829 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.172899 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.171074 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166235 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166375 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166447 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166593 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166752 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166952 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.167332 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.167498 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.166197 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.171447 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.171432 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.171464 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.172126 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.172304 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.172579 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.172600 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.172715 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.172837 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.172867 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.172989 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:55:20.672947762 +0000 UTC m=+20.573539507 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.176094 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.176129 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.176147 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.176145 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.176189 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.173319 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.173186 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.173486 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.174352 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.174512 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.174530 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.174818 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.174956 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.175029 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.175154 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.175418 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.176351 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.176421 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.176444 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.177935 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.176783 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.177022 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.177745 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.176165 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178065 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178089 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178109 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178129 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178154 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178177 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178199 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178220 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178241 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178260 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178281 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178303 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178323 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178322 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178344 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178374 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178420 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178449 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178473 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178495 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178517 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178592 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178603 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178618 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178644 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178666 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178688 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178713 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178734 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178755 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178779 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178842 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178854 4809 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178867 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178883 4809 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178921 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178936 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178948 4809 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178961 4809 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.178990 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179004 4809 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179016 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179028 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179028 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179043 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179058 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179070 4809 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179081 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179093 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179106 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179118 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179130 4809 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179141 4809 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179154 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179167 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179182 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179195 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179206 4809 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179219 4809 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179231 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179243 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179256 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179267 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179282 4809 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179293 4809 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179304 4809 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179316 4809 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179327 4809 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179339 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179351 4809 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179365 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179378 4809 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179453 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179467 4809 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179478 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179489 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179500 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179511 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179523 4809 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179535 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179547 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179559 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179570 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179582 4809 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179597 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179608 4809 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179620 4809 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179631 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179644 4809 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179655 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179666 4809 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179678 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179688 4809 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179699 4809 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179709 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179725 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179736 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179748 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179759 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179770 4809 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179781 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179793 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179804 4809 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179816 4809 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179829 4809 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179842 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179852 4809 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179864 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179875 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179886 4809 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179897 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179907 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179917 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179928 4809 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179939 4809 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179950 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179979 4809 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.179991 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180002 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180012 4809 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180023 4809 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180034 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180045 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180058 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180069 4809 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180080 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180091 4809 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180107 4809 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180123 4809 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180135 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180147 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180158 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180170 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180181 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180192 4809 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180203 4809 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180213 4809 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180224 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180235 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180246 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180257 4809 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180268 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180279 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180289 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180300 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180310 4809 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180321 4809 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180332 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180347 4809 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180359 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180371 4809 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.180382 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.180825 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.180874 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.180944 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:20.680864669 +0000 UTC m=+20.581456374 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.180958 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:20.680952561 +0000 UTC m=+20.581544266 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.173190 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.181148 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.181585 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.181904 4809 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.182325 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.182617 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.182658 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.182683 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.183071 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.183272 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.184510 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.184563 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.185400 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.186179 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.186215 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.186303 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.186365 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.187132 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.187154 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.187181 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.187393 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.187550 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.187614 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.187673 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.191164 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.191190 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.191725 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.192315 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.193058 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.193244 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.190385 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.198248 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.204653 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.205631 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.206043 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.210416 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.210795 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.211953 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.211997 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.212034 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.212429 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.212469 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.212550 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.212580 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.212646 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.212757 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:20.71273256 +0000 UTC m=+20.613324265 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.212843 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.212870 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.212884 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.212925 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:20.712909676 +0000 UTC m=+20.613501381 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.213124 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.214779 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.215554 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.215946 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.216307 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.216440 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.216772 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.217066 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.217345 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.219362 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.219482 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.219937 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.220228 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.220535 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.220607 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.220628 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.225327 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.226749 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.232018 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.232074 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.232438 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.233192 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.234770 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.235513 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.236765 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.236837 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.238524 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.243980 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.244154 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.244622 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.244637 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.244661 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.244794 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.244911 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.245160 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.245396 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.246549 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.247712 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.259701 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.267281 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281335 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281386 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281421 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281432 4809 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281441 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281449 4809 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281457 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281465 4809 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281473 4809 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281481 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281489 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281496 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281504 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281512 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281520 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281528 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281536 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281544 4809 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281552 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281560 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281568 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281577 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281585 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281594 4809 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281603 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281610 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281618 4809 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281626 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281661 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281670 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281679 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281687 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281699 4809 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281707 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281715 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281723 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281730 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281738 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281748 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281765 4809 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281779 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281788 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281798 4809 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281808 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281818 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281826 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281835 4809 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281844 4809 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281853 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281860 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281868 4809 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281875 4809 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281883 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281891 4809 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281899 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281907 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281915 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281922 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281930 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281939 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281947 4809 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281956 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.281987 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.282003 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.282014 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.282025 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.282036 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.282047 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.282057 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.282068 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.282079 4809 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.282089 4809 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.282097 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.282130 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.282153 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.282162 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.282202 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.282332 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.289696 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-8p7tz"] Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.290186 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-8p7tz" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.293094 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.293307 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.293459 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.312887 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.333387 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.341745 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.349994 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.355476 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.361801 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.368498 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.382809 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62n55\" (UniqueName: \"kubernetes.io/projected/3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09-kube-api-access-62n55\") pod \"node-resolver-8p7tz\" (UID: \"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\") " pod="openshift-dns/node-resolver-8p7tz" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.382874 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09-hosts-file\") pod \"node-resolver-8p7tz\" (UID: \"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\") " pod="openshift-dns/node-resolver-8p7tz" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.458278 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 06:55:20 crc kubenswrapper[4809]: W1124 06:55:20.469933 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-bf698289cb4645057a5a35d693e09748018dd44cdca8ffa65a2b99effb9da39a WatchSource:0}: Error finding container bf698289cb4645057a5a35d693e09748018dd44cdca8ffa65a2b99effb9da39a: Status 404 returned error can't find the container with id bf698289cb4645057a5a35d693e09748018dd44cdca8ffa65a2b99effb9da39a Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.475689 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.483690 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09-hosts-file\") pod \"node-resolver-8p7tz\" (UID: \"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\") " pod="openshift-dns/node-resolver-8p7tz" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.483739 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62n55\" (UniqueName: \"kubernetes.io/projected/3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09-kube-api-access-62n55\") pod \"node-resolver-8p7tz\" (UID: \"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\") " pod="openshift-dns/node-resolver-8p7tz" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.483915 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09-hosts-file\") pod \"node-resolver-8p7tz\" (UID: \"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\") " pod="openshift-dns/node-resolver-8p7tz" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.484548 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 06:55:20 crc kubenswrapper[4809]: W1124 06:55:20.499371 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-4a2d69b6c6adc2e0071406912549a64cc0db35bbc93f9143c61f40424c83785e WatchSource:0}: Error finding container 4a2d69b6c6adc2e0071406912549a64cc0db35bbc93f9143c61f40424c83785e: Status 404 returned error can't find the container with id 4a2d69b6c6adc2e0071406912549a64cc0db35bbc93f9143c61f40424c83785e Nov 24 06:55:20 crc kubenswrapper[4809]: W1124 06:55:20.499770 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-83b164093b3c498bae07aa1616b71ebaeefee91da891fd186ecceec9e41ad974 WatchSource:0}: Error finding container 83b164093b3c498bae07aa1616b71ebaeefee91da891fd186ecceec9e41ad974: Status 404 returned error can't find the container with id 83b164093b3c498bae07aa1616b71ebaeefee91da891fd186ecceec9e41ad974 Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.504255 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62n55\" (UniqueName: \"kubernetes.io/projected/3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09-kube-api-access-62n55\") pod \"node-resolver-8p7tz\" (UID: \"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\") " pod="openshift-dns/node-resolver-8p7tz" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.613645 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-8p7tz" Nov 24 06:55:20 crc kubenswrapper[4809]: W1124 06:55:20.623460 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3cc8cb5c_3973_4f32_984b_5e1c7b4e9f09.slice/crio-2578dc347957b2662bf3fc0d6bd5ab78cc579a9371ddf974ff260de029efda81 WatchSource:0}: Error finding container 2578dc347957b2662bf3fc0d6bd5ab78cc579a9371ddf974ff260de029efda81: Status 404 returned error can't find the container with id 2578dc347957b2662bf3fc0d6bd5ab78cc579a9371ddf974ff260de029efda81 Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.633028 4809 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Nov 24 06:55:20 crc kubenswrapper[4809]: W1124 06:55:20.633201 4809 reflector.go:484] object-"openshift-dns"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Nov 24 06:55:20 crc kubenswrapper[4809]: W1124 06:55:20.633246 4809 reflector.go:484] object-"openshift-network-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Nov 24 06:55:20 crc kubenswrapper[4809]: W1124 06:55:20.633270 4809 reflector.go:484] object-"openshift-network-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Nov 24 06:55:20 crc kubenswrapper[4809]: W1124 06:55:20.633294 4809 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.CSIDriver ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Nov 24 06:55:20 crc kubenswrapper[4809]: W1124 06:55:20.633328 4809 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.RuntimeClass ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Nov 24 06:55:20 crc kubenswrapper[4809]: W1124 06:55:20.633348 4809 reflector.go:484] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": watch of *v1.Secret ended with: very short watch: object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": Unexpected watch close - watch lasted less than a second and no items received Nov 24 06:55:20 crc kubenswrapper[4809]: W1124 06:55:20.634132 4809 reflector.go:484] object-"openshift-network-node-identity"/"ovnkube-identity-cm": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"ovnkube-identity-cm": Unexpected watch close - watch lasted less than a second and no items received Nov 24 06:55:20 crc kubenswrapper[4809]: W1124 06:55:20.634189 4809 reflector.go:484] object-"openshift-network-node-identity"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Nov 24 06:55:20 crc kubenswrapper[4809]: W1124 06:55:20.634219 4809 reflector.go:484] object-"openshift-network-operator"/"metrics-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-network-operator"/"metrics-tls": Unexpected watch close - watch lasted less than a second and no items received Nov 24 06:55:20 crc kubenswrapper[4809]: W1124 06:55:20.634239 4809 reflector.go:484] object-"openshift-network-operator"/"iptables-alerter-script": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"iptables-alerter-script": Unexpected watch close - watch lasted less than a second and no items received Nov 24 06:55:20 crc kubenswrapper[4809]: W1124 06:55:20.634256 4809 reflector.go:484] object-"openshift-network-node-identity"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Nov 24 06:55:20 crc kubenswrapper[4809]: W1124 06:55:20.634243 4809 reflector.go:484] object-"openshift-dns"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Nov 24 06:55:20 crc kubenswrapper[4809]: W1124 06:55:20.634268 4809 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.Service ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Nov 24 06:55:20 crc kubenswrapper[4809]: W1124 06:55:20.634293 4809 reflector.go:484] object-"openshift-network-node-identity"/"network-node-identity-cert": watch of *v1.Secret ended with: very short watch: object-"openshift-network-node-identity"/"network-node-identity-cert": Unexpected watch close - watch lasted less than a second and no items received Nov 24 06:55:20 crc kubenswrapper[4809]: W1124 06:55:20.634201 4809 reflector.go:484] object-"openshift-network-node-identity"/"env-overrides": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"env-overrides": Unexpected watch close - watch lasted less than a second and no items received Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.649640 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.656635 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.659080 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.667946 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.680438 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.684593 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.684695 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:55:21.684670522 +0000 UTC m=+21.585262237 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.689013 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.689051 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.689131 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.689185 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:21.68916909 +0000 UTC m=+21.589760795 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.689256 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.689286 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:21.689276903 +0000 UTC m=+21.589868618 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.700779 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.712219 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.719541 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.727458 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.735167 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.748794 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.762558 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.772316 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.781586 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.789746 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.789790 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.789920 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.789940 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.789951 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.790014 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:21.789996896 +0000 UTC m=+21.690588601 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.790076 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.790090 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.790099 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.790126 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:21.79011717 +0000 UTC m=+21.690708875 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.790530 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.803431 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.813371 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.822530 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.890611 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:20 crc kubenswrapper[4809]: E1124 06:55:20.890735 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.894639 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.895477 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.896334 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.897049 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.897670 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.899077 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.899744 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.900801 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.901468 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.902425 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.902947 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.904088 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.904612 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.905218 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.906247 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.906808 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.908074 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.908644 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.909101 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.909743 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.911102 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.911674 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.912702 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.914005 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.915130 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.915613 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.916265 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.918430 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.919374 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.921372 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.922156 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.922268 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.924790 4809 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.925031 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.927044 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.928586 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.930426 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.933421 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.933780 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.935155 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.936542 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.937296 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.939259 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.939851 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.941024 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.941724 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.943261 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.944071 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.944776 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.945344 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.946061 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.947249 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.947812 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.948705 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.949510 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.950605 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.951231 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.951765 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.954063 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.963084 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.977591 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.979194 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.981241 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2" exitCode=255 Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.981366 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2"} Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.983196 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-8p7tz" event={"ID":"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09","Type":"ContainerStarted","Data":"50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6"} Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.983256 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-8p7tz" event={"ID":"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09","Type":"ContainerStarted","Data":"2578dc347957b2662bf3fc0d6bd5ab78cc579a9371ddf974ff260de029efda81"} Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.985836 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d"} Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.985874 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"bf698289cb4645057a5a35d693e09748018dd44cdca8ffa65a2b99effb9da39a"} Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.985956 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"83b164093b3c498bae07aa1616b71ebaeefee91da891fd186ecceec9e41ad974"} Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.988205 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c"} Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.988226 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446"} Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.988237 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4a2d69b6c6adc2e0071406912549a64cc0db35bbc93f9143c61f40424c83785e"} Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.988939 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.992827 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 24 06:55:20 crc kubenswrapper[4809]: I1124 06:55:20.992907 4809 scope.go:117] "RemoveContainer" containerID="854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.003058 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.013813 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.028133 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.040367 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.048926 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.058364 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.068667 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.078302 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.108032 4809 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-11-24 06:50:20 +0000 UTC, rotation deadline is 2026-09-27 14:48:57.640906793 +0000 UTC Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.108118 4809 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7375h53m36.532791809s for next certificate rotation Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.478297 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.500503 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.518712 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.564296 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.696887 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:55:21 crc kubenswrapper[4809]: E1124 06:55:21.697188 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:55:23.697161346 +0000 UTC m=+23.597753051 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.697278 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.697322 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:21 crc kubenswrapper[4809]: E1124 06:55:21.697404 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:55:21 crc kubenswrapper[4809]: E1124 06:55:21.697439 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:23.697432484 +0000 UTC m=+23.598024189 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:55:21 crc kubenswrapper[4809]: E1124 06:55:21.697581 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:55:21 crc kubenswrapper[4809]: E1124 06:55:21.697697 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:23.697680301 +0000 UTC m=+23.598272006 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.705201 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.751025 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.781094 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.798528 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.798580 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:21 crc kubenswrapper[4809]: E1124 06:55:21.798735 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:55:21 crc kubenswrapper[4809]: E1124 06:55:21.798756 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:55:21 crc kubenswrapper[4809]: E1124 06:55:21.798771 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:21 crc kubenswrapper[4809]: E1124 06:55:21.798821 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:55:21 crc kubenswrapper[4809]: E1124 06:55:21.798867 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:55:21 crc kubenswrapper[4809]: E1124 06:55:21.798886 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:21 crc kubenswrapper[4809]: E1124 06:55:21.798829 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:23.798813777 +0000 UTC m=+23.699405492 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:21 crc kubenswrapper[4809]: E1124 06:55:21.799027 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:23.798990302 +0000 UTC m=+23.699582007 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.833514 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.882204 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.890776 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.890789 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:21 crc kubenswrapper[4809]: E1124 06:55:21.890907 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:55:21 crc kubenswrapper[4809]: E1124 06:55:21.891027 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.978429 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.983440 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.992491 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.994076 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d"} Nov 24 06:55:21 crc kubenswrapper[4809]: I1124 06:55:21.994586 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.006789 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.023814 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.031665 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.043167 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.060789 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.077289 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.092065 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.118091 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.131441 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-dr8hv"] Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.131841 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.132899 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-kpqhl"] Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.133458 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.134471 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-5xvl8"] Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.134665 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.134907 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.135358 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.135387 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.135748 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.135790 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.135999 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.136024 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.136216 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.136308 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.136343 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.136613 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-8wxqg"] Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.137224 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.137537 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.140110 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.140114 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.140114 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.140116 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.140401 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.140440 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.140454 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.140519 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.144354 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.163993 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.182262 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.196953 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202427 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-hostroot\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202463 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-etc-openvswitch\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202479 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/74622ad8-44b0-44a7-afbd-e19bb9635520-ovnkube-config\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202502 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-host-run-netns\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202516 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-host-var-lib-cni-bin\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202529 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-host-var-lib-cni-multus\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202544 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sl4qj\" (UniqueName: \"kubernetes.io/projected/be8050d5-1aea-453e-a06e-87702b0e856b-kube-api-access-sl4qj\") pod \"machine-config-daemon-dr8hv\" (UID: \"be8050d5-1aea-453e-a06e-87702b0e856b\") " pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202559 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-multus-socket-dir-parent\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202572 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-multus-conf-dir\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202586 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-run-systemd\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202600 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-etc-kubernetes\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202613 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gs7tj\" (UniqueName: \"kubernetes.io/projected/3c932894-56b4-4303-8176-d12b12686b09-kube-api-access-gs7tj\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202627 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-multus-cni-dir\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202640 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-var-lib-openvswitch\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202654 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-run-openvswitch\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202668 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-host-run-multus-certs\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202681 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-host-var-lib-kubelet\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202695 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3c932894-56b4-4303-8176-d12b12686b09-cni-binary-copy\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202709 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202725 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7hjg\" (UniqueName: \"kubernetes.io/projected/74622ad8-44b0-44a7-afbd-e19bb9635520-kube-api-access-t7hjg\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202739 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/74622ad8-44b0-44a7-afbd-e19bb9635520-ovnkube-script-lib\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202752 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-cni-bin\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202765 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/74622ad8-44b0-44a7-afbd-e19bb9635520-env-overrides\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202791 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-system-cni-dir\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202804 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-kubelet\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202820 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-run-ovn-kubernetes\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202834 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-os-release\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202847 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-node-log\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202861 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/be8050d5-1aea-453e-a06e-87702b0e856b-rootfs\") pod \"machine-config-daemon-dr8hv\" (UID: \"be8050d5-1aea-453e-a06e-87702b0e856b\") " pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202877 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/be8050d5-1aea-453e-a06e-87702b0e856b-proxy-tls\") pod \"machine-config-daemon-dr8hv\" (UID: \"be8050d5-1aea-453e-a06e-87702b0e856b\") " pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202892 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/be8050d5-1aea-453e-a06e-87702b0e856b-mcd-auth-proxy-config\") pod \"machine-config-daemon-dr8hv\" (UID: \"be8050d5-1aea-453e-a06e-87702b0e856b\") " pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202913 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-log-socket\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202926 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-cni-netd\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202947 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-run-ovn\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.202980 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/74622ad8-44b0-44a7-afbd-e19bb9635520-ovn-node-metrics-cert\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.203001 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-slash\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.203020 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-cnibin\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.203038 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-systemd-units\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.203059 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-host-run-k8s-cni-cncf-io\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.203077 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-run-netns\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.203199 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/3c932894-56b4-4303-8176-d12b12686b09-multus-daemon-config\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.211279 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.222798 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.232997 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.244235 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.256255 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.267649 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.279472 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.295583 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.303657 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-node-log\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.303696 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-os-release\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.303714 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/be8050d5-1aea-453e-a06e-87702b0e856b-rootfs\") pod \"machine-config-daemon-dr8hv\" (UID: \"be8050d5-1aea-453e-a06e-87702b0e856b\") " pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.303731 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/be8050d5-1aea-453e-a06e-87702b0e856b-proxy-tls\") pod \"machine-config-daemon-dr8hv\" (UID: \"be8050d5-1aea-453e-a06e-87702b0e856b\") " pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.303750 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/be8050d5-1aea-453e-a06e-87702b0e856b-mcd-auth-proxy-config\") pod \"machine-config-daemon-dr8hv\" (UID: \"be8050d5-1aea-453e-a06e-87702b0e856b\") " pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.303771 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-log-socket\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.303785 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-cni-netd\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.303804 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-run-ovn\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.303807 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-node-log\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.303828 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/be8050d5-1aea-453e-a06e-87702b0e856b-rootfs\") pod \"machine-config-daemon-dr8hv\" (UID: \"be8050d5-1aea-453e-a06e-87702b0e856b\") " pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.303882 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-cni-netd\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.303818 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/74622ad8-44b0-44a7-afbd-e19bb9635520-ovn-node-metrics-cert\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.303930 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-systemd-units\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.303952 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-slash\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.303991 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/86405460-8db2-444c-8273-f2e7040733d2-tuning-conf-dir\") pod \"multus-additional-cni-plugins-8wxqg\" (UID: \"86405460-8db2-444c-8273-f2e7040733d2\") " pod="openshift-multus/multus-additional-cni-plugins-8wxqg" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304017 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-cnibin\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304071 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-host-run-k8s-cni-cncf-io\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304092 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/3c932894-56b4-4303-8176-d12b12686b09-multus-daemon-config\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304112 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-run-netns\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304133 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/86405460-8db2-444c-8273-f2e7040733d2-cnibin\") pod \"multus-additional-cni-plugins-8wxqg\" (UID: \"86405460-8db2-444c-8273-f2e7040733d2\") " pod="openshift-multus/multus-additional-cni-plugins-8wxqg" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304167 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-host-var-lib-cni-multus\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304185 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-hostroot\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304201 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-etc-openvswitch\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304219 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/74622ad8-44b0-44a7-afbd-e19bb9635520-ovnkube-config\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304238 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/86405460-8db2-444c-8273-f2e7040733d2-system-cni-dir\") pod \"multus-additional-cni-plugins-8wxqg\" (UID: \"86405460-8db2-444c-8273-f2e7040733d2\") " pod="openshift-multus/multus-additional-cni-plugins-8wxqg" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304265 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-host-run-netns\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304283 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-host-var-lib-cni-bin\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304304 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4mkk\" (UniqueName: \"kubernetes.io/projected/86405460-8db2-444c-8273-f2e7040733d2-kube-api-access-r4mkk\") pod \"multus-additional-cni-plugins-8wxqg\" (UID: \"86405460-8db2-444c-8273-f2e7040733d2\") " pod="openshift-multus/multus-additional-cni-plugins-8wxqg" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304323 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-multus-socket-dir-parent\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304342 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sl4qj\" (UniqueName: \"kubernetes.io/projected/be8050d5-1aea-453e-a06e-87702b0e856b-kube-api-access-sl4qj\") pod \"machine-config-daemon-dr8hv\" (UID: \"be8050d5-1aea-453e-a06e-87702b0e856b\") " pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304365 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-multus-conf-dir\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304382 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-run-systemd\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304399 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-etc-kubernetes\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304419 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gs7tj\" (UniqueName: \"kubernetes.io/projected/3c932894-56b4-4303-8176-d12b12686b09-kube-api-access-gs7tj\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304439 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/86405460-8db2-444c-8273-f2e7040733d2-cni-binary-copy\") pod \"multus-additional-cni-plugins-8wxqg\" (UID: \"86405460-8db2-444c-8273-f2e7040733d2\") " pod="openshift-multus/multus-additional-cni-plugins-8wxqg" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304457 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/86405460-8db2-444c-8273-f2e7040733d2-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-8wxqg\" (UID: \"86405460-8db2-444c-8273-f2e7040733d2\") " pod="openshift-multus/multus-additional-cni-plugins-8wxqg" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304477 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-multus-cni-dir\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304501 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-var-lib-openvswitch\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304520 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-host-run-multus-certs\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304538 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-run-openvswitch\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304559 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-host-var-lib-kubelet\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304581 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7hjg\" (UniqueName: \"kubernetes.io/projected/74622ad8-44b0-44a7-afbd-e19bb9635520-kube-api-access-t7hjg\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304603 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/86405460-8db2-444c-8273-f2e7040733d2-os-release\") pod \"multus-additional-cni-plugins-8wxqg\" (UID: \"86405460-8db2-444c-8273-f2e7040733d2\") " pod="openshift-multus/multus-additional-cni-plugins-8wxqg" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304625 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3c932894-56b4-4303-8176-d12b12686b09-cni-binary-copy\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304645 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304669 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/74622ad8-44b0-44a7-afbd-e19bb9635520-ovnkube-script-lib\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304692 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-run-ovn-kubernetes\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304714 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-cni-bin\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304736 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/74622ad8-44b0-44a7-afbd-e19bb9635520-env-overrides\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304736 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-run-ovn\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304757 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-system-cni-dir\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304779 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-kubelet\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304841 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-kubelet\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304877 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-multus-conf-dir\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304908 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-run-systemd\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304936 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-etc-kubernetes\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304958 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-log-socket\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.304985 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/be8050d5-1aea-453e-a06e-87702b0e856b-mcd-auth-proxy-config\") pod \"machine-config-daemon-dr8hv\" (UID: \"be8050d5-1aea-453e-a06e-87702b0e856b\") " pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305003 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-systemd-units\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305030 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-slash\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305047 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-hostroot\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305054 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-etc-openvswitch\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305116 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-cnibin\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305148 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-host-run-k8s-cni-cncf-io\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305151 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-os-release\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305189 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-run-netns\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305227 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-host-var-lib-cni-multus\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305250 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-multus-cni-dir\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305258 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-host-var-lib-cni-bin\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305289 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-host-run-netns\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305315 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-var-lib-openvswitch\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305319 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-host-var-lib-kubelet\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305335 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-host-run-multus-certs\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305358 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-run-openvswitch\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305560 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/74622ad8-44b0-44a7-afbd-e19bb9635520-ovnkube-config\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305565 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-multus-socket-dir-parent\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305599 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-run-ovn-kubernetes\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305624 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305860 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/3c932894-56b4-4303-8176-d12b12686b09-multus-daemon-config\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305926 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3c932894-56b4-4303-8176-d12b12686b09-system-cni-dir\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305938 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3c932894-56b4-4303-8176-d12b12686b09-cni-binary-copy\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.305975 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-cni-bin\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.306045 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/74622ad8-44b0-44a7-afbd-e19bb9635520-env-overrides\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.306173 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/74622ad8-44b0-44a7-afbd-e19bb9635520-ovnkube-script-lib\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.309428 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/74622ad8-44b0-44a7-afbd-e19bb9635520-ovn-node-metrics-cert\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.309433 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/be8050d5-1aea-453e-a06e-87702b0e856b-proxy-tls\") pod \"machine-config-daemon-dr8hv\" (UID: \"be8050d5-1aea-453e-a06e-87702b0e856b\") " pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.328054 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gs7tj\" (UniqueName: \"kubernetes.io/projected/3c932894-56b4-4303-8176-d12b12686b09-kube-api-access-gs7tj\") pod \"multus-5xvl8\" (UID: \"3c932894-56b4-4303-8176-d12b12686b09\") " pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.332587 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sl4qj\" (UniqueName: \"kubernetes.io/projected/be8050d5-1aea-453e-a06e-87702b0e856b-kube-api-access-sl4qj\") pod \"machine-config-daemon-dr8hv\" (UID: \"be8050d5-1aea-453e-a06e-87702b0e856b\") " pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.343616 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7hjg\" (UniqueName: \"kubernetes.io/projected/74622ad8-44b0-44a7-afbd-e19bb9635520-kube-api-access-t7hjg\") pod \"ovnkube-node-kpqhl\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.348983 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.361920 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.384118 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.405406 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/86405460-8db2-444c-8273-f2e7040733d2-tuning-conf-dir\") pod \"multus-additional-cni-plugins-8wxqg\" (UID: \"86405460-8db2-444c-8273-f2e7040733d2\") " pod="openshift-multus/multus-additional-cni-plugins-8wxqg" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.405445 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/86405460-8db2-444c-8273-f2e7040733d2-cnibin\") pod \"multus-additional-cni-plugins-8wxqg\" (UID: \"86405460-8db2-444c-8273-f2e7040733d2\") " pod="openshift-multus/multus-additional-cni-plugins-8wxqg" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.405468 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/86405460-8db2-444c-8273-f2e7040733d2-system-cni-dir\") pod \"multus-additional-cni-plugins-8wxqg\" (UID: \"86405460-8db2-444c-8273-f2e7040733d2\") " pod="openshift-multus/multus-additional-cni-plugins-8wxqg" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.405488 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4mkk\" (UniqueName: \"kubernetes.io/projected/86405460-8db2-444c-8273-f2e7040733d2-kube-api-access-r4mkk\") pod \"multus-additional-cni-plugins-8wxqg\" (UID: \"86405460-8db2-444c-8273-f2e7040733d2\") " pod="openshift-multus/multus-additional-cni-plugins-8wxqg" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.405507 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/86405460-8db2-444c-8273-f2e7040733d2-cni-binary-copy\") pod \"multus-additional-cni-plugins-8wxqg\" (UID: \"86405460-8db2-444c-8273-f2e7040733d2\") " pod="openshift-multus/multus-additional-cni-plugins-8wxqg" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.405521 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/86405460-8db2-444c-8273-f2e7040733d2-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-8wxqg\" (UID: \"86405460-8db2-444c-8273-f2e7040733d2\") " pod="openshift-multus/multus-additional-cni-plugins-8wxqg" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.405539 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/86405460-8db2-444c-8273-f2e7040733d2-os-release\") pod \"multus-additional-cni-plugins-8wxqg\" (UID: \"86405460-8db2-444c-8273-f2e7040733d2\") " pod="openshift-multus/multus-additional-cni-plugins-8wxqg" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.405599 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/86405460-8db2-444c-8273-f2e7040733d2-system-cni-dir\") pod \"multus-additional-cni-plugins-8wxqg\" (UID: \"86405460-8db2-444c-8273-f2e7040733d2\") " pod="openshift-multus/multus-additional-cni-plugins-8wxqg" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.405622 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/86405460-8db2-444c-8273-f2e7040733d2-os-release\") pod \"multus-additional-cni-plugins-8wxqg\" (UID: \"86405460-8db2-444c-8273-f2e7040733d2\") " pod="openshift-multus/multus-additional-cni-plugins-8wxqg" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.406181 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/86405460-8db2-444c-8273-f2e7040733d2-tuning-conf-dir\") pod \"multus-additional-cni-plugins-8wxqg\" (UID: \"86405460-8db2-444c-8273-f2e7040733d2\") " pod="openshift-multus/multus-additional-cni-plugins-8wxqg" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.406373 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/86405460-8db2-444c-8273-f2e7040733d2-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-8wxqg\" (UID: \"86405460-8db2-444c-8273-f2e7040733d2\") " pod="openshift-multus/multus-additional-cni-plugins-8wxqg" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.406395 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/86405460-8db2-444c-8273-f2e7040733d2-cni-binary-copy\") pod \"multus-additional-cni-plugins-8wxqg\" (UID: \"86405460-8db2-444c-8273-f2e7040733d2\") " pod="openshift-multus/multus-additional-cni-plugins-8wxqg" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.406461 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/86405460-8db2-444c-8273-f2e7040733d2-cnibin\") pod \"multus-additional-cni-plugins-8wxqg\" (UID: \"86405460-8db2-444c-8273-f2e7040733d2\") " pod="openshift-multus/multus-additional-cni-plugins-8wxqg" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.425906 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4mkk\" (UniqueName: \"kubernetes.io/projected/86405460-8db2-444c-8273-f2e7040733d2-kube-api-access-r4mkk\") pod \"multus-additional-cni-plugins-8wxqg\" (UID: \"86405460-8db2-444c-8273-f2e7040733d2\") " pod="openshift-multus/multus-additional-cni-plugins-8wxqg" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.449249 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.452264 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.459529 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-5xvl8" Nov 24 06:55:22 crc kubenswrapper[4809]: W1124 06:55:22.465592 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74622ad8_44b0_44a7_afbd_e19bb9635520.slice/crio-be0abc73293942a2e9751f120dda9b69eb20a33702af201e3c4e611d18e7705f WatchSource:0}: Error finding container be0abc73293942a2e9751f120dda9b69eb20a33702af201e3c4e611d18e7705f: Status 404 returned error can't find the container with id be0abc73293942a2e9751f120dda9b69eb20a33702af201e3c4e611d18e7705f Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.466824 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" Nov 24 06:55:22 crc kubenswrapper[4809]: W1124 06:55:22.469021 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe8050d5_1aea_453e_a06e_87702b0e856b.slice/crio-04454ec00177f73c0b4fd6d616161b5d2a67d097c656d167ca7ef19e7b7440da WatchSource:0}: Error finding container 04454ec00177f73c0b4fd6d616161b5d2a67d097c656d167ca7ef19e7b7440da: Status 404 returned error can't find the container with id 04454ec00177f73c0b4fd6d616161b5d2a67d097c656d167ca7ef19e7b7440da Nov 24 06:55:22 crc kubenswrapper[4809]: W1124 06:55:22.479254 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86405460_8db2_444c_8273_f2e7040733d2.slice/crio-8f7a321e1b1d6fcb8331ceab871eaa1f647c18566309056a94958a1c578126e9 WatchSource:0}: Error finding container 8f7a321e1b1d6fcb8331ceab871eaa1f647c18566309056a94958a1c578126e9: Status 404 returned error can't find the container with id 8f7a321e1b1d6fcb8331ceab871eaa1f647c18566309056a94958a1c578126e9 Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.890625 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:22 crc kubenswrapper[4809]: E1124 06:55:22.891013 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:55:22 crc kubenswrapper[4809]: I1124 06:55:22.998229 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675"} Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.000273 4809 generic.go:334] "Generic (PLEG): container finished" podID="86405460-8db2-444c-8273-f2e7040733d2" containerID="209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1" exitCode=0 Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.000370 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" event={"ID":"86405460-8db2-444c-8273-f2e7040733d2","Type":"ContainerDied","Data":"209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1"} Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.000445 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" event={"ID":"86405460-8db2-444c-8273-f2e7040733d2","Type":"ContainerStarted","Data":"8f7a321e1b1d6fcb8331ceab871eaa1f647c18566309056a94958a1c578126e9"} Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.002070 4809 generic.go:334] "Generic (PLEG): container finished" podID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerID="9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855" exitCode=0 Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.002122 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerDied","Data":"9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855"} Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.002139 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerStarted","Data":"be0abc73293942a2e9751f120dda9b69eb20a33702af201e3c4e611d18e7705f"} Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.004175 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5xvl8" event={"ID":"3c932894-56b4-4303-8176-d12b12686b09","Type":"ContainerStarted","Data":"0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95"} Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.004202 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5xvl8" event={"ID":"3c932894-56b4-4303-8176-d12b12686b09","Type":"ContainerStarted","Data":"983884fed6e5b709a8aa38ebb319995e7a01ab7079472e5e4ced48036befcad7"} Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.008514 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerStarted","Data":"07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e"} Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.010983 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerStarted","Data":"a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5"} Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.010999 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerStarted","Data":"04454ec00177f73c0b4fd6d616161b5d2a67d097c656d167ca7ef19e7b7440da"} Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.018551 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.037229 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.050350 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.067554 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.083527 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.099604 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.111827 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.126815 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.141001 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.153817 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.167206 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.181501 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.201325 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.218890 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.230979 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.247556 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.260176 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.275721 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.287584 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.303592 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.317586 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.329375 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.339315 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.352403 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.364747 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.377807 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.718876 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:55:23 crc kubenswrapper[4809]: E1124 06:55:23.719104 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:55:27.719075559 +0000 UTC m=+27.619667274 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.719273 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.719296 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:23 crc kubenswrapper[4809]: E1124 06:55:23.719393 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:55:23 crc kubenswrapper[4809]: E1124 06:55:23.719393 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:55:23 crc kubenswrapper[4809]: E1124 06:55:23.719445 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:27.719431639 +0000 UTC m=+27.620023344 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:55:23 crc kubenswrapper[4809]: E1124 06:55:23.719459 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:27.71945321 +0000 UTC m=+27.620044905 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.820486 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:23 crc kubenswrapper[4809]: E1124 06:55:23.820688 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:55:23 crc kubenswrapper[4809]: E1124 06:55:23.820859 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:55:23 crc kubenswrapper[4809]: E1124 06:55:23.820874 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:23 crc kubenswrapper[4809]: E1124 06:55:23.820927 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.820804 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:23 crc kubenswrapper[4809]: E1124 06:55:23.820950 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:55:23 crc kubenswrapper[4809]: E1124 06:55:23.821100 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:23 crc kubenswrapper[4809]: E1124 06:55:23.820932 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:27.820912694 +0000 UTC m=+27.721504409 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:23 crc kubenswrapper[4809]: E1124 06:55:23.821205 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:27.821176382 +0000 UTC m=+27.721768117 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.891049 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:23 crc kubenswrapper[4809]: I1124 06:55:23.891060 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:23 crc kubenswrapper[4809]: E1124 06:55:23.891172 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:55:23 crc kubenswrapper[4809]: E1124 06:55:23.891299 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:55:24 crc kubenswrapper[4809]: I1124 06:55:24.017830 4809 generic.go:334] "Generic (PLEG): container finished" podID="86405460-8db2-444c-8273-f2e7040733d2" containerID="44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f" exitCode=0 Nov 24 06:55:24 crc kubenswrapper[4809]: I1124 06:55:24.017949 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" event={"ID":"86405460-8db2-444c-8273-f2e7040733d2","Type":"ContainerDied","Data":"44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f"} Nov 24 06:55:24 crc kubenswrapper[4809]: I1124 06:55:24.026292 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerStarted","Data":"787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a"} Nov 24 06:55:24 crc kubenswrapper[4809]: I1124 06:55:24.026374 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerStarted","Data":"cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844"} Nov 24 06:55:24 crc kubenswrapper[4809]: I1124 06:55:24.026406 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerStarted","Data":"0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844"} Nov 24 06:55:24 crc kubenswrapper[4809]: I1124 06:55:24.026433 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerStarted","Data":"88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249"} Nov 24 06:55:24 crc kubenswrapper[4809]: I1124 06:55:24.026463 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerStarted","Data":"54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed"} Nov 24 06:55:24 crc kubenswrapper[4809]: I1124 06:55:24.034838 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:24 crc kubenswrapper[4809]: I1124 06:55:24.047183 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:24 crc kubenswrapper[4809]: I1124 06:55:24.058755 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:24 crc kubenswrapper[4809]: I1124 06:55:24.076651 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:24 crc kubenswrapper[4809]: I1124 06:55:24.094758 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:24 crc kubenswrapper[4809]: I1124 06:55:24.108874 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:24 crc kubenswrapper[4809]: I1124 06:55:24.124778 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:24 crc kubenswrapper[4809]: I1124 06:55:24.135586 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:24 crc kubenswrapper[4809]: I1124 06:55:24.147874 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:24 crc kubenswrapper[4809]: I1124 06:55:24.167007 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:24 crc kubenswrapper[4809]: I1124 06:55:24.180947 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:24 crc kubenswrapper[4809]: I1124 06:55:24.193368 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:24 crc kubenswrapper[4809]: I1124 06:55:24.204723 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:24 crc kubenswrapper[4809]: I1124 06:55:24.890261 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:24 crc kubenswrapper[4809]: E1124 06:55:24.890445 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:55:25 crc kubenswrapper[4809]: I1124 06:55:25.032276 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerStarted","Data":"62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7"} Nov 24 06:55:25 crc kubenswrapper[4809]: I1124 06:55:25.034663 4809 generic.go:334] "Generic (PLEG): container finished" podID="86405460-8db2-444c-8273-f2e7040733d2" containerID="fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b" exitCode=0 Nov 24 06:55:25 crc kubenswrapper[4809]: I1124 06:55:25.034690 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" event={"ID":"86405460-8db2-444c-8273-f2e7040733d2","Type":"ContainerDied","Data":"fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b"} Nov 24 06:55:25 crc kubenswrapper[4809]: I1124 06:55:25.066988 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:25 crc kubenswrapper[4809]: I1124 06:55:25.081576 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:25 crc kubenswrapper[4809]: I1124 06:55:25.096062 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:25 crc kubenswrapper[4809]: I1124 06:55:25.109854 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:25 crc kubenswrapper[4809]: I1124 06:55:25.122334 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:25 crc kubenswrapper[4809]: I1124 06:55:25.133121 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:25 crc kubenswrapper[4809]: I1124 06:55:25.146258 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:25 crc kubenswrapper[4809]: I1124 06:55:25.163179 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:25 crc kubenswrapper[4809]: I1124 06:55:25.176415 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:25 crc kubenswrapper[4809]: I1124 06:55:25.187107 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:25 crc kubenswrapper[4809]: I1124 06:55:25.199662 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:25 crc kubenswrapper[4809]: I1124 06:55:25.211899 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:25 crc kubenswrapper[4809]: I1124 06:55:25.228003 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:25 crc kubenswrapper[4809]: I1124 06:55:25.765470 4809 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 24 06:55:25 crc kubenswrapper[4809]: I1124 06:55:25.890090 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:25 crc kubenswrapper[4809]: E1124 06:55:25.890449 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:55:25 crc kubenswrapper[4809]: I1124 06:55:25.890090 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:25 crc kubenswrapper[4809]: E1124 06:55:25.890632 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.040265 4809 generic.go:334] "Generic (PLEG): container finished" podID="86405460-8db2-444c-8273-f2e7040733d2" containerID="795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58" exitCode=0 Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.040333 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" event={"ID":"86405460-8db2-444c-8273-f2e7040733d2","Type":"ContainerDied","Data":"795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58"} Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.080074 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:26Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.096567 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:26Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.112245 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:26Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.156452 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:26Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.169410 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:26Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.183672 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:26Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.200046 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:26Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.212680 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:26Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.222604 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:26Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.235826 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:26Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.247790 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:26Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.259800 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:26Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.276222 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:26Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.278774 4809 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.452459 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.455680 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.455742 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.455766 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.455895 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.463868 4809 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.464138 4809 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.465394 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.465460 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.465474 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.465499 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.465513 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:26Z","lastTransitionTime":"2025-11-24T06:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:26 crc kubenswrapper[4809]: E1124 06:55:26.479162 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:26Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.483668 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.483698 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.483709 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.483724 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.483735 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:26Z","lastTransitionTime":"2025-11-24T06:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:26 crc kubenswrapper[4809]: E1124 06:55:26.497319 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:26Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.501218 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.501250 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.501262 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.501278 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.501287 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:26Z","lastTransitionTime":"2025-11-24T06:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:26 crc kubenswrapper[4809]: E1124 06:55:26.515257 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:26Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.518478 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.518527 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.518544 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.518566 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.518581 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:26Z","lastTransitionTime":"2025-11-24T06:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:26 crc kubenswrapper[4809]: E1124 06:55:26.532628 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:26Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.536388 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.536417 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.536426 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.536440 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.536454 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:26Z","lastTransitionTime":"2025-11-24T06:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:26 crc kubenswrapper[4809]: E1124 06:55:26.549264 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:26Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:26 crc kubenswrapper[4809]: E1124 06:55:26.549375 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.551445 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.551479 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.551492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.551509 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.551520 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:26Z","lastTransitionTime":"2025-11-24T06:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.653375 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.653411 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.653421 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.653466 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.653479 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:26Z","lastTransitionTime":"2025-11-24T06:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.755190 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.755243 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.755253 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.755285 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.755294 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:26Z","lastTransitionTime":"2025-11-24T06:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.857357 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.857454 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.857472 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.857495 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.857512 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:26Z","lastTransitionTime":"2025-11-24T06:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.891196 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:26 crc kubenswrapper[4809]: E1124 06:55:26.891380 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.920178 4809 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.959523 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.959562 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.959578 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.959595 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:26 crc kubenswrapper[4809]: I1124 06:55:26.959605 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:26Z","lastTransitionTime":"2025-11-24T06:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.048137 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerStarted","Data":"a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c"} Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.054444 4809 generic.go:334] "Generic (PLEG): container finished" podID="86405460-8db2-444c-8273-f2e7040733d2" containerID="5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b" exitCode=0 Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.054496 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" event={"ID":"86405460-8db2-444c-8273-f2e7040733d2","Type":"ContainerDied","Data":"5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b"} Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.061260 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.061305 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.061317 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.061334 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.061345 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:27Z","lastTransitionTime":"2025-11-24T06:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.079850 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:27Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.099581 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:27Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.118173 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:27Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.129009 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:27Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.142801 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:27Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.155875 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:27Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.164329 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.164359 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.164369 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.164384 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.164430 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:27Z","lastTransitionTime":"2025-11-24T06:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.167828 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:27Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.179010 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:27Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.190295 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:27Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.203372 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:27Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.215164 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:27Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.224434 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:27Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.235990 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:27Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.266930 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.267013 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.267030 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.267048 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.267061 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:27Z","lastTransitionTime":"2025-11-24T06:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.369071 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.369101 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.369111 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.369123 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.369132 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:27Z","lastTransitionTime":"2025-11-24T06:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.472069 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.472106 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.472115 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.472129 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.472137 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:27Z","lastTransitionTime":"2025-11-24T06:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.574639 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.574678 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.574687 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.574701 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.574710 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:27Z","lastTransitionTime":"2025-11-24T06:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.677821 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.677858 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.677866 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.677882 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.677890 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:27Z","lastTransitionTime":"2025-11-24T06:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.771292 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.771413 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.771446 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:27 crc kubenswrapper[4809]: E1124 06:55:27.771497 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:55:35.771471091 +0000 UTC m=+35.672062796 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:55:27 crc kubenswrapper[4809]: E1124 06:55:27.771540 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:55:27 crc kubenswrapper[4809]: E1124 06:55:27.771607 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:35.771591724 +0000 UTC m=+35.672183509 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:55:27 crc kubenswrapper[4809]: E1124 06:55:27.771671 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:55:27 crc kubenswrapper[4809]: E1124 06:55:27.771778 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:35.771759399 +0000 UTC m=+35.672351144 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.781661 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.781719 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.781731 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.781759 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.781773 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:27Z","lastTransitionTime":"2025-11-24T06:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.872079 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.872127 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:27 crc kubenswrapper[4809]: E1124 06:55:27.872238 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:55:27 crc kubenswrapper[4809]: E1124 06:55:27.872254 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:55:27 crc kubenswrapper[4809]: E1124 06:55:27.872264 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:27 crc kubenswrapper[4809]: E1124 06:55:27.872317 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:35.872304727 +0000 UTC m=+35.772896432 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:27 crc kubenswrapper[4809]: E1124 06:55:27.872428 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:55:27 crc kubenswrapper[4809]: E1124 06:55:27.872509 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:55:27 crc kubenswrapper[4809]: E1124 06:55:27.872533 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:27 crc kubenswrapper[4809]: E1124 06:55:27.872636 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:35.872602816 +0000 UTC m=+35.773194711 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.884472 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.884519 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.884530 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.884547 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.884560 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:27Z","lastTransitionTime":"2025-11-24T06:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.891206 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:27 crc kubenswrapper[4809]: E1124 06:55:27.891316 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.891324 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:27 crc kubenswrapper[4809]: E1124 06:55:27.891556 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.987419 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.987457 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.987471 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.987490 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:27 crc kubenswrapper[4809]: I1124 06:55:27.987500 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:27Z","lastTransitionTime":"2025-11-24T06:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.063780 4809 generic.go:334] "Generic (PLEG): container finished" podID="86405460-8db2-444c-8273-f2e7040733d2" containerID="5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d" exitCode=0 Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.063859 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" event={"ID":"86405460-8db2-444c-8273-f2e7040733d2","Type":"ContainerDied","Data":"5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d"} Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.088265 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:28Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.090026 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.090070 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.090084 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.090107 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.090121 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:28Z","lastTransitionTime":"2025-11-24T06:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.103259 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:28Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.125365 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:28Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.143545 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:28Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.160947 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:28Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.177521 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:28Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.191351 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:28Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.192506 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.192526 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.192533 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.192545 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.192554 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:28Z","lastTransitionTime":"2025-11-24T06:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.202445 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:28Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.213949 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:28Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.224423 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:28Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.234487 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:28Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.246263 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:28Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.259414 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:28Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.294673 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.294723 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.294734 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.294751 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.294761 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:28Z","lastTransitionTime":"2025-11-24T06:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.401119 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.401153 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.401164 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.401179 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.401190 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:28Z","lastTransitionTime":"2025-11-24T06:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.504285 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.504715 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.504724 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.504742 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.504754 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:28Z","lastTransitionTime":"2025-11-24T06:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.615813 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.615864 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.615877 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.615902 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.615915 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:28Z","lastTransitionTime":"2025-11-24T06:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.719119 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.719196 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.719209 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.719254 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.719268 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:28Z","lastTransitionTime":"2025-11-24T06:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.822590 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.822637 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.822650 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.822676 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.822690 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:28Z","lastTransitionTime":"2025-11-24T06:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.890810 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:28 crc kubenswrapper[4809]: E1124 06:55:28.890936 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.925780 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.925815 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.925825 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.925839 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:28 crc kubenswrapper[4809]: I1124 06:55:28.925851 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:28Z","lastTransitionTime":"2025-11-24T06:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.028565 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.028654 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.028678 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.028705 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.028727 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:29Z","lastTransitionTime":"2025-11-24T06:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.073673 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" event={"ID":"86405460-8db2-444c-8273-f2e7040733d2","Type":"ContainerStarted","Data":"f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541"} Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.086103 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerStarted","Data":"b121b0c3597cfaab37db70e8e7d7533d88b8aaf9a9f8a0e6e3e3e92786129c4a"} Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.086674 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.087984 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.100245 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.110053 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.123100 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.129009 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.130892 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.131023 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.131052 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.131130 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.131158 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:29Z","lastTransitionTime":"2025-11-24T06:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.145691 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.158477 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.171708 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.186487 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.198170 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.207994 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.221724 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.233528 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.233658 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.233698 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.233710 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.233724 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.233734 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:29Z","lastTransitionTime":"2025-11-24T06:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.245138 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.257584 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.268480 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.278343 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.291005 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.312436 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b121b0c3597cfaab37db70e8e7d7533d88b8aaf9a9f8a0e6e3e3e92786129c4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.327658 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.336402 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.336450 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.336463 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.336481 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.336493 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:29Z","lastTransitionTime":"2025-11-24T06:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.341320 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.352383 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.363898 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.373064 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.388676 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.400760 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.413392 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.438871 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.438908 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.438917 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.438931 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.438940 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:29Z","lastTransitionTime":"2025-11-24T06:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.541184 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.541223 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.541232 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.541244 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.541254 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:29Z","lastTransitionTime":"2025-11-24T06:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.643224 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.643252 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.643262 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.643279 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.643292 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:29Z","lastTransitionTime":"2025-11-24T06:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.745510 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.745547 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.745556 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.745570 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.745578 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:29Z","lastTransitionTime":"2025-11-24T06:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.848588 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.848628 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.848642 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.848659 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.848672 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:29Z","lastTransitionTime":"2025-11-24T06:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.891236 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.891275 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:29 crc kubenswrapper[4809]: E1124 06:55:29.891356 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:55:29 crc kubenswrapper[4809]: E1124 06:55:29.891423 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.950512 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.950551 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.950560 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.950574 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:29 crc kubenswrapper[4809]: I1124 06:55:29.950583 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:29Z","lastTransitionTime":"2025-11-24T06:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.053437 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.053471 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.053480 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.053494 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.053503 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:30Z","lastTransitionTime":"2025-11-24T06:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.090050 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.090727 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.114526 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.126322 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:30Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.137249 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:30Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.146855 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:30Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.155221 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.155250 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.155257 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.155270 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.155282 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:30Z","lastTransitionTime":"2025-11-24T06:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.158792 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:30Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.174096 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b121b0c3597cfaab37db70e8e7d7533d88b8aaf9a9f8a0e6e3e3e92786129c4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:30Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.183605 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:30Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.196505 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:30Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.209951 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:30Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.223237 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:30Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.239319 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:30Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.254882 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:30Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.258100 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.258193 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.258206 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.258221 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.258230 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:30Z","lastTransitionTime":"2025-11-24T06:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.269132 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:30Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.279783 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:30Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.362035 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.362091 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.362110 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.362134 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.362154 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:30Z","lastTransitionTime":"2025-11-24T06:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.465456 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.465504 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.465515 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.465534 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.465546 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:30Z","lastTransitionTime":"2025-11-24T06:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.567512 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.567576 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.567585 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.567598 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.567607 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:30Z","lastTransitionTime":"2025-11-24T06:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.669668 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.669695 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.669703 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.669717 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.669725 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:30Z","lastTransitionTime":"2025-11-24T06:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.771764 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.771794 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.771802 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.771815 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.771824 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:30Z","lastTransitionTime":"2025-11-24T06:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.873824 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.873856 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.873864 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.873877 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.873908 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:30Z","lastTransitionTime":"2025-11-24T06:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.894532 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:30 crc kubenswrapper[4809]: E1124 06:55:30.894755 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.929804 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b121b0c3597cfaab37db70e8e7d7533d88b8aaf9a9f8a0e6e3e3e92786129c4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:30Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.945651 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:30Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.955778 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:30Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.965823 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:30Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.975426 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.975468 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.975481 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.975500 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.975512 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:30Z","lastTransitionTime":"2025-11-24T06:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:30 crc kubenswrapper[4809]: I1124 06:55:30.985300 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:30Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:30.999878 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:30Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.014955 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:31Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.031989 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:31Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.043108 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:31Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.057548 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:31Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.068741 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:31Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.077896 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.077950 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.077987 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.078008 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.078020 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:31Z","lastTransitionTime":"2025-11-24T06:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.080280 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:31Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.091376 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:31Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.092596 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.180449 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.180483 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.180492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.180505 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.180514 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:31Z","lastTransitionTime":"2025-11-24T06:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.282884 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.282927 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.282939 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.282954 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.282981 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:31Z","lastTransitionTime":"2025-11-24T06:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.385495 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.385786 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.385797 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.385812 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.385822 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:31Z","lastTransitionTime":"2025-11-24T06:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.488104 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.488153 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.488167 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.488187 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.488204 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:31Z","lastTransitionTime":"2025-11-24T06:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.590311 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.590353 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.590364 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.590379 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.590390 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:31Z","lastTransitionTime":"2025-11-24T06:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.693069 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.693112 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.693120 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.693134 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.693144 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:31Z","lastTransitionTime":"2025-11-24T06:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.795601 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.795631 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.795640 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.795653 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.795663 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:31Z","lastTransitionTime":"2025-11-24T06:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.891142 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.891157 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:31 crc kubenswrapper[4809]: E1124 06:55:31.891262 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:55:31 crc kubenswrapper[4809]: E1124 06:55:31.891348 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.897627 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.897664 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.897672 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.897686 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:31 crc kubenswrapper[4809]: I1124 06:55:31.897696 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:31Z","lastTransitionTime":"2025-11-24T06:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.000610 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.000657 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.000668 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.000684 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.000696 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:32Z","lastTransitionTime":"2025-11-24T06:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.096588 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kpqhl_74622ad8-44b0-44a7-afbd-e19bb9635520/ovnkube-controller/0.log" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.099205 4809 generic.go:334] "Generic (PLEG): container finished" podID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerID="b121b0c3597cfaab37db70e8e7d7533d88b8aaf9a9f8a0e6e3e3e92786129c4a" exitCode=1 Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.099245 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerDied","Data":"b121b0c3597cfaab37db70e8e7d7533d88b8aaf9a9f8a0e6e3e3e92786129c4a"} Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.099899 4809 scope.go:117] "RemoveContainer" containerID="b121b0c3597cfaab37db70e8e7d7533d88b8aaf9a9f8a0e6e3e3e92786129c4a" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.103084 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.103112 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.103123 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.103138 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.103149 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:32Z","lastTransitionTime":"2025-11-24T06:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.117812 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b121b0c3597cfaab37db70e8e7d7533d88b8aaf9a9f8a0e6e3e3e92786129c4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b121b0c3597cfaab37db70e8e7d7533d88b8aaf9a9f8a0e6e3e3e92786129c4a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:31Z\\\",\\\"message\\\":\\\"\\\\nI1124 06:55:31.450533 6091 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:55:31.450612 6091 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 06:55:31.450687 6091 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 06:55:31.450835 6091 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 06:55:31.450853 6091 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 06:55:31.450864 6091 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 06:55:31.450889 6091 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 06:55:31.450924 6091 factory.go:656] Stopping watch factory\\\\nI1124 06:55:31.450946 6091 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 06:55:31.450958 6091 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 06:55:31.450980 6091 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 06:55:31.450989 6091 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.129325 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.140680 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.152619 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.163460 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.174178 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.184489 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.196171 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.205433 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.205466 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.205474 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.205486 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.205496 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:32Z","lastTransitionTime":"2025-11-24T06:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.210478 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.223611 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.240888 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.254525 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.267535 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.307110 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.307142 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.307151 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.307164 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.307174 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:32Z","lastTransitionTime":"2025-11-24T06:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.410292 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.410343 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.410354 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.410371 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.410382 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:32Z","lastTransitionTime":"2025-11-24T06:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.512368 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.512408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.512416 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.512433 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.512442 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:32Z","lastTransitionTime":"2025-11-24T06:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.614609 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.614645 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.614653 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.614668 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.614680 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:32Z","lastTransitionTime":"2025-11-24T06:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.716722 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.716763 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.716774 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.716790 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.716802 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:32Z","lastTransitionTime":"2025-11-24T06:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.819040 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.819090 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.819111 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.819131 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.819144 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:32Z","lastTransitionTime":"2025-11-24T06:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.891086 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:32 crc kubenswrapper[4809]: E1124 06:55:32.891221 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.921354 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.921395 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.921404 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.921419 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:32 crc kubenswrapper[4809]: I1124 06:55:32.921429 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:32Z","lastTransitionTime":"2025-11-24T06:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.023662 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.023706 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.023718 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.023735 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.023747 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:33Z","lastTransitionTime":"2025-11-24T06:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.065405 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.076715 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.088805 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.103926 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.104571 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kpqhl_74622ad8-44b0-44a7-afbd-e19bb9635520/ovnkube-controller/1.log" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.105056 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kpqhl_74622ad8-44b0-44a7-afbd-e19bb9635520/ovnkube-controller/0.log" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.111198 4809 generic.go:334] "Generic (PLEG): container finished" podID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerID="743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40" exitCode=1 Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.111254 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerDied","Data":"743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40"} Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.111299 4809 scope.go:117] "RemoveContainer" containerID="b121b0c3597cfaab37db70e8e7d7533d88b8aaf9a9f8a0e6e3e3e92786129c4a" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.111991 4809 scope.go:117] "RemoveContainer" containerID="743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40" Nov 24 06:55:33 crc kubenswrapper[4809]: E1124 06:55:33.112145 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-kpqhl_openshift-ovn-kubernetes(74622ad8-44b0-44a7-afbd-e19bb9635520)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.119476 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.125528 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.125565 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.125577 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.125595 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.125607 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:33Z","lastTransitionTime":"2025-11-24T06:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.129957 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.141801 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.155468 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.167126 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.177641 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.189083 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.202404 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.214725 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.228359 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.228402 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.228413 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.228430 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.228442 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:33Z","lastTransitionTime":"2025-11-24T06:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.232545 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b121b0c3597cfaab37db70e8e7d7533d88b8aaf9a9f8a0e6e3e3e92786129c4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b121b0c3597cfaab37db70e8e7d7533d88b8aaf9a9f8a0e6e3e3e92786129c4a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:31Z\\\",\\\"message\\\":\\\"\\\\nI1124 06:55:31.450533 6091 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:55:31.450612 6091 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 06:55:31.450687 6091 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 06:55:31.450835 6091 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 06:55:31.450853 6091 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 06:55:31.450864 6091 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 06:55:31.450889 6091 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 06:55:31.450924 6091 factory.go:656] Stopping watch factory\\\\nI1124 06:55:31.450946 6091 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 06:55:31.450958 6091 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 06:55:31.450980 6091 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 06:55:31.450989 6091 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.246062 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.258994 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.271946 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.282061 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.292207 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.304924 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.316659 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.327302 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.330789 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.330819 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.330828 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.330841 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.330850 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:33Z","lastTransitionTime":"2025-11-24T06:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.338922 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.348650 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.368724 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.387754 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.415984 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b121b0c3597cfaab37db70e8e7d7533d88b8aaf9a9f8a0e6e3e3e92786129c4a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:31Z\\\",\\\"message\\\":\\\"\\\\nI1124 06:55:31.450533 6091 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:55:31.450612 6091 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 06:55:31.450687 6091 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 06:55:31.450835 6091 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 06:55:31.450853 6091 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 06:55:31.450864 6091 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 06:55:31.450889 6091 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 06:55:31.450924 6091 factory.go:656] Stopping watch factory\\\\nI1124 06:55:31.450946 6091 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 06:55:31.450958 6091 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 06:55:31.450980 6091 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 06:55:31.450989 6091 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:32Z\\\",\\\"message\\\":\\\"rce:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8444, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1124 06:55:32.828572 6221 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z]\\\\nI1124 06:55:32.828645 6221 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manage\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:33Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.432374 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.432422 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.432437 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.432458 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.432479 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:33Z","lastTransitionTime":"2025-11-24T06:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.534843 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.534882 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.534891 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.534905 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.534914 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:33Z","lastTransitionTime":"2025-11-24T06:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.636555 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.636604 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.636612 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.636626 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.636634 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:33Z","lastTransitionTime":"2025-11-24T06:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.738825 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.738858 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.738866 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.738878 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.738886 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:33Z","lastTransitionTime":"2025-11-24T06:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.841591 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.841624 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.841632 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.841644 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.841652 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:33Z","lastTransitionTime":"2025-11-24T06:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.890710 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.890710 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:33 crc kubenswrapper[4809]: E1124 06:55:33.890841 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:55:33 crc kubenswrapper[4809]: E1124 06:55:33.891006 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.976129 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.976163 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.976173 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.976188 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:33 crc kubenswrapper[4809]: I1124 06:55:33.976199 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:33Z","lastTransitionTime":"2025-11-24T06:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.078897 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.078948 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.078960 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.079003 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.079016 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:34Z","lastTransitionTime":"2025-11-24T06:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.115488 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kpqhl_74622ad8-44b0-44a7-afbd-e19bb9635520/ovnkube-controller/1.log" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.118766 4809 scope.go:117] "RemoveContainer" containerID="743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40" Nov 24 06:55:34 crc kubenswrapper[4809]: E1124 06:55:34.118899 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-kpqhl_openshift-ovn-kubernetes(74622ad8-44b0-44a7-afbd-e19bb9635520)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.141740 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:32Z\\\",\\\"message\\\":\\\"rce:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8444, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1124 06:55:32.828572 6221 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z]\\\\nI1124 06:55:32.828645 6221 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manage\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-kpqhl_openshift-ovn-kubernetes(74622ad8-44b0-44a7-afbd-e19bb9635520)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.154117 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.165601 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.172311 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr"] Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.172998 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.173675 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-7cq8b"] Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.174110 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7cq8b" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.177342 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.178622 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.178636 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.178808 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.179017 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.180319 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.182357 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.182396 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.182408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.182425 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.182438 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:34Z","lastTransitionTime":"2025-11-24T06:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.186448 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.200847 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.213429 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.226806 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.235942 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.249696 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.262864 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.274753 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.284868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.284898 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.284909 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.284925 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.284938 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:34Z","lastTransitionTime":"2025-11-24T06:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.291203 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.306522 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.318426 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.328134 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.336136 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbdhx\" (UniqueName: \"kubernetes.io/projected/15c4f8f9-fd68-4636-88cb-d967cb3e202c-kube-api-access-wbdhx\") pod \"node-ca-7cq8b\" (UID: \"15c4f8f9-fd68-4636-88cb-d967cb3e202c\") " pod="openshift-image-registry/node-ca-7cq8b" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.336212 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0d459b25-158a-4a2f-bff7-b7c643450c10-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-5fvqr\" (UID: \"0d459b25-158a-4a2f-bff7-b7c643450c10\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.336242 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/15c4f8f9-fd68-4636-88cb-d967cb3e202c-host\") pod \"node-ca-7cq8b\" (UID: \"15c4f8f9-fd68-4636-88cb-d967cb3e202c\") " pod="openshift-image-registry/node-ca-7cq8b" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.336274 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/15c4f8f9-fd68-4636-88cb-d967cb3e202c-serviceca\") pod \"node-ca-7cq8b\" (UID: \"15c4f8f9-fd68-4636-88cb-d967cb3e202c\") " pod="openshift-image-registry/node-ca-7cq8b" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.336296 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtv8n\" (UniqueName: \"kubernetes.io/projected/0d459b25-158a-4a2f-bff7-b7c643450c10-kube-api-access-mtv8n\") pod \"ovnkube-control-plane-749d76644c-5fvqr\" (UID: \"0d459b25-158a-4a2f-bff7-b7c643450c10\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.336386 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0d459b25-158a-4a2f-bff7-b7c643450c10-env-overrides\") pod \"ovnkube-control-plane-749d76644c-5fvqr\" (UID: \"0d459b25-158a-4a2f-bff7-b7c643450c10\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.336430 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0d459b25-158a-4a2f-bff7-b7c643450c10-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-5fvqr\" (UID: \"0d459b25-158a-4a2f-bff7-b7c643450c10\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.338802 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.351492 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.361638 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d459b25-158a-4a2f-bff7-b7c643450c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5fvqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.379266 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:32Z\\\",\\\"message\\\":\\\"rce:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8444, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1124 06:55:32.828572 6221 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z]\\\\nI1124 06:55:32.828645 6221 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manage\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-kpqhl_openshift-ovn-kubernetes(74622ad8-44b0-44a7-afbd-e19bb9635520)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.387493 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.387551 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.387571 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.387593 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.387610 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:34Z","lastTransitionTime":"2025-11-24T06:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.393639 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cq8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15c4f8f9-fd68-4636-88cb-d967cb3e202c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbdhx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cq8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.404864 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.417030 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.431058 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.437791 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/15c4f8f9-fd68-4636-88cb-d967cb3e202c-serviceca\") pod \"node-ca-7cq8b\" (UID: \"15c4f8f9-fd68-4636-88cb-d967cb3e202c\") " pod="openshift-image-registry/node-ca-7cq8b" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.437892 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtv8n\" (UniqueName: \"kubernetes.io/projected/0d459b25-158a-4a2f-bff7-b7c643450c10-kube-api-access-mtv8n\") pod \"ovnkube-control-plane-749d76644c-5fvqr\" (UID: \"0d459b25-158a-4a2f-bff7-b7c643450c10\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.438012 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0d459b25-158a-4a2f-bff7-b7c643450c10-env-overrides\") pod \"ovnkube-control-plane-749d76644c-5fvqr\" (UID: \"0d459b25-158a-4a2f-bff7-b7c643450c10\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.438072 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0d459b25-158a-4a2f-bff7-b7c643450c10-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-5fvqr\" (UID: \"0d459b25-158a-4a2f-bff7-b7c643450c10\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.438141 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbdhx\" (UniqueName: \"kubernetes.io/projected/15c4f8f9-fd68-4636-88cb-d967cb3e202c-kube-api-access-wbdhx\") pod \"node-ca-7cq8b\" (UID: \"15c4f8f9-fd68-4636-88cb-d967cb3e202c\") " pod="openshift-image-registry/node-ca-7cq8b" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.438246 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0d459b25-158a-4a2f-bff7-b7c643450c10-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-5fvqr\" (UID: \"0d459b25-158a-4a2f-bff7-b7c643450c10\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.438292 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/15c4f8f9-fd68-4636-88cb-d967cb3e202c-host\") pod \"node-ca-7cq8b\" (UID: \"15c4f8f9-fd68-4636-88cb-d967cb3e202c\") " pod="openshift-image-registry/node-ca-7cq8b" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.438399 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/15c4f8f9-fd68-4636-88cb-d967cb3e202c-host\") pod \"node-ca-7cq8b\" (UID: \"15c4f8f9-fd68-4636-88cb-d967cb3e202c\") " pod="openshift-image-registry/node-ca-7cq8b" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.438960 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/15c4f8f9-fd68-4636-88cb-d967cb3e202c-serviceca\") pod \"node-ca-7cq8b\" (UID: \"15c4f8f9-fd68-4636-88cb-d967cb3e202c\") " pod="openshift-image-registry/node-ca-7cq8b" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.439100 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0d459b25-158a-4a2f-bff7-b7c643450c10-env-overrides\") pod \"ovnkube-control-plane-749d76644c-5fvqr\" (UID: \"0d459b25-158a-4a2f-bff7-b7c643450c10\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.439697 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0d459b25-158a-4a2f-bff7-b7c643450c10-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-5fvqr\" (UID: \"0d459b25-158a-4a2f-bff7-b7c643450c10\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.444391 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.450917 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0d459b25-158a-4a2f-bff7-b7c643450c10-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-5fvqr\" (UID: \"0d459b25-158a-4a2f-bff7-b7c643450c10\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.453681 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbdhx\" (UniqueName: \"kubernetes.io/projected/15c4f8f9-fd68-4636-88cb-d967cb3e202c-kube-api-access-wbdhx\") pod \"node-ca-7cq8b\" (UID: \"15c4f8f9-fd68-4636-88cb-d967cb3e202c\") " pod="openshift-image-registry/node-ca-7cq8b" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.454748 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.465288 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.471037 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtv8n\" (UniqueName: \"kubernetes.io/projected/0d459b25-158a-4a2f-bff7-b7c643450c10-kube-api-access-mtv8n\") pod \"ovnkube-control-plane-749d76644c-5fvqr\" (UID: \"0d459b25-158a-4a2f-bff7-b7c643450c10\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.476464 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.485776 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.489667 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.489726 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.489741 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.489764 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.489779 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:34Z","lastTransitionTime":"2025-11-24T06:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.494802 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.499367 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7cq8b" Nov 24 06:55:34 crc kubenswrapper[4809]: W1124 06:55:34.511332 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d459b25_158a_4a2f_bff7_b7c643450c10.slice/crio-1b833230216d43d797e86e3fc2442da5470f94188912a1f2695f78a7b2ebed6a WatchSource:0}: Error finding container 1b833230216d43d797e86e3fc2442da5470f94188912a1f2695f78a7b2ebed6a: Status 404 returned error can't find the container with id 1b833230216d43d797e86e3fc2442da5470f94188912a1f2695f78a7b2ebed6a Nov 24 06:55:34 crc kubenswrapper[4809]: W1124 06:55:34.512495 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15c4f8f9_fd68_4636_88cb_d967cb3e202c.slice/crio-7ad3f018d9dd287d9d911623142d2938471864bd68fef0e9420ad4d66cbed376 WatchSource:0}: Error finding container 7ad3f018d9dd287d9d911623142d2938471864bd68fef0e9420ad4d66cbed376: Status 404 returned error can't find the container with id 7ad3f018d9dd287d9d911623142d2938471864bd68fef0e9420ad4d66cbed376 Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.525018 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-krmz8"] Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.525476 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:55:34 crc kubenswrapper[4809]: E1124 06:55:34.525530 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.538723 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.550561 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.560283 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.574307 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.585257 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d459b25-158a-4a2f-bff7-b7c643450c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5fvqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.591755 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.591787 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.591798 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.591815 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.591827 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:34Z","lastTransitionTime":"2025-11-24T06:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.594399 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cq8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15c4f8f9-fd68-4636-88cb-d967cb3e202c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbdhx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cq8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.610175 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:32Z\\\",\\\"message\\\":\\\"rce:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8444, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1124 06:55:32.828572 6221 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z]\\\\nI1124 06:55:32.828645 6221 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manage\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-kpqhl_openshift-ovn-kubernetes(74622ad8-44b0-44a7-afbd-e19bb9635520)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.622683 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.634460 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.639783 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76zdl\" (UniqueName: \"kubernetes.io/projected/35b8615e-faa5-47b4-8c22-609d5fda9590-kube-api-access-76zdl\") pod \"network-metrics-daemon-krmz8\" (UID: \"35b8615e-faa5-47b4-8c22-609d5fda9590\") " pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.639837 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs\") pod \"network-metrics-daemon-krmz8\" (UID: \"35b8615e-faa5-47b4-8c22-609d5fda9590\") " pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.643274 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.653008 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.665546 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.677879 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.689752 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.694917 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.694955 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.694986 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.695002 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.695014 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:34Z","lastTransitionTime":"2025-11-24T06:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.703781 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.712760 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krmz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b8615e-faa5-47b4-8c22-609d5fda9590\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krmz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:34Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.741424 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs\") pod \"network-metrics-daemon-krmz8\" (UID: \"35b8615e-faa5-47b4-8c22-609d5fda9590\") " pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.741522 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76zdl\" (UniqueName: \"kubernetes.io/projected/35b8615e-faa5-47b4-8c22-609d5fda9590-kube-api-access-76zdl\") pod \"network-metrics-daemon-krmz8\" (UID: \"35b8615e-faa5-47b4-8c22-609d5fda9590\") " pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:55:34 crc kubenswrapper[4809]: E1124 06:55:34.741753 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:55:34 crc kubenswrapper[4809]: E1124 06:55:34.741827 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs podName:35b8615e-faa5-47b4-8c22-609d5fda9590 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:35.241811405 +0000 UTC m=+35.142403110 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs") pod "network-metrics-daemon-krmz8" (UID: "35b8615e-faa5-47b4-8c22-609d5fda9590") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.759985 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76zdl\" (UniqueName: \"kubernetes.io/projected/35b8615e-faa5-47b4-8c22-609d5fda9590-kube-api-access-76zdl\") pod \"network-metrics-daemon-krmz8\" (UID: \"35b8615e-faa5-47b4-8c22-609d5fda9590\") " pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.797693 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.797724 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.797732 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.797745 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.797754 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:34Z","lastTransitionTime":"2025-11-24T06:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.890952 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:34 crc kubenswrapper[4809]: E1124 06:55:34.891109 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.899484 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.899512 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.899522 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.899537 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:34 crc kubenswrapper[4809]: I1124 06:55:34.899547 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:34Z","lastTransitionTime":"2025-11-24T06:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.001600 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.001646 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.001658 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.001674 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.001684 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:35Z","lastTransitionTime":"2025-11-24T06:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.104461 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.104535 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.104559 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.104592 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.104614 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:35Z","lastTransitionTime":"2025-11-24T06:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.122897 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" event={"ID":"0d459b25-158a-4a2f-bff7-b7c643450c10","Type":"ContainerStarted","Data":"bd69b51611cd90f67fb36b564db5890aff54f52a1884d71d297f335d707ec8c2"} Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.123229 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" event={"ID":"0d459b25-158a-4a2f-bff7-b7c643450c10","Type":"ContainerStarted","Data":"fc5172a4b40f579c4ae113095afcdc882c281d31c0ddf2f59f0593c0a8e1837c"} Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.123303 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" event={"ID":"0d459b25-158a-4a2f-bff7-b7c643450c10","Type":"ContainerStarted","Data":"1b833230216d43d797e86e3fc2442da5470f94188912a1f2695f78a7b2ebed6a"} Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.124855 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7cq8b" event={"ID":"15c4f8f9-fd68-4636-88cb-d967cb3e202c","Type":"ContainerStarted","Data":"2d9fa21c921ed12c61aa6a10a4cd4a5c575109a77a6b7099106b76414358d1e1"} Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.124901 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7cq8b" event={"ID":"15c4f8f9-fd68-4636-88cb-d967cb3e202c","Type":"ContainerStarted","Data":"7ad3f018d9dd287d9d911623142d2938471864bd68fef0e9420ad4d66cbed376"} Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.142716 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:32Z\\\",\\\"message\\\":\\\"rce:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8444, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1124 06:55:32.828572 6221 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z]\\\\nI1124 06:55:32.828645 6221 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manage\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-kpqhl_openshift-ovn-kubernetes(74622ad8-44b0-44a7-afbd-e19bb9635520)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.154932 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cq8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15c4f8f9-fd68-4636-88cb-d967cb3e202c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbdhx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cq8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.167383 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.186991 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.199113 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.207553 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.207597 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.207608 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.207621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.207630 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:35Z","lastTransitionTime":"2025-11-24T06:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.210713 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.220632 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.234760 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.245325 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.245624 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs\") pod \"network-metrics-daemon-krmz8\" (UID: \"35b8615e-faa5-47b4-8c22-609d5fda9590\") " pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:55:35 crc kubenswrapper[4809]: E1124 06:55:35.245732 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:55:35 crc kubenswrapper[4809]: E1124 06:55:35.245796 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs podName:35b8615e-faa5-47b4-8c22-609d5fda9590 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:36.245783233 +0000 UTC m=+36.146374938 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs") pod "network-metrics-daemon-krmz8" (UID: "35b8615e-faa5-47b4-8c22-609d5fda9590") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.257373 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.270784 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krmz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b8615e-faa5-47b4-8c22-609d5fda9590\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krmz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.287435 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.300099 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.315113 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.315167 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.315179 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.315196 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.315207 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:35Z","lastTransitionTime":"2025-11-24T06:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.318330 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.334252 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.346164 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d459b25-158a-4a2f-bff7-b7c643450c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc5172a4b40f579c4ae113095afcdc882c281d31c0ddf2f59f0593c0a8e1837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69b51611cd90f67fb36b564db5890aff54f52a1884d71d297f335d707ec8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5fvqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.356685 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d459b25-158a-4a2f-bff7-b7c643450c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc5172a4b40f579c4ae113095afcdc882c281d31c0ddf2f59f0593c0a8e1837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69b51611cd90f67fb36b564db5890aff54f52a1884d71d297f335d707ec8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5fvqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.377828 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:32Z\\\",\\\"message\\\":\\\"rce:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8444, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1124 06:55:32.828572 6221 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z]\\\\nI1124 06:55:32.828645 6221 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manage\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-kpqhl_openshift-ovn-kubernetes(74622ad8-44b0-44a7-afbd-e19bb9635520)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.388951 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cq8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15c4f8f9-fd68-4636-88cb-d967cb3e202c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d9fa21c921ed12c61aa6a10a4cd4a5c575109a77a6b7099106b76414358d1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbdhx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cq8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.398632 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.409400 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.418446 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.418495 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.418509 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.418530 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.418543 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:35Z","lastTransitionTime":"2025-11-24T06:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.423846 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.441169 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.458168 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.472415 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.484035 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.496106 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.506690 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krmz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b8615e-faa5-47b4-8c22-609d5fda9590\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krmz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.517017 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.520148 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.520173 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.520184 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.520200 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.520211 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:35Z","lastTransitionTime":"2025-11-24T06:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.528347 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.537735 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.550344 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.622490 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.622542 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.622554 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.622571 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.622583 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:35Z","lastTransitionTime":"2025-11-24T06:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.724912 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.724954 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.724986 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.725006 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.725020 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:35Z","lastTransitionTime":"2025-11-24T06:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.827486 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.827546 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.827561 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.827579 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.827593 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:35Z","lastTransitionTime":"2025-11-24T06:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.850864 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:55:35 crc kubenswrapper[4809]: E1124 06:55:35.850959 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:55:51.850940347 +0000 UTC m=+51.751532052 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.851051 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.851083 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:35 crc kubenswrapper[4809]: E1124 06:55:35.851309 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:55:35 crc kubenswrapper[4809]: E1124 06:55:35.851414 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:51.85140122 +0000 UTC m=+51.751992925 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:55:35 crc kubenswrapper[4809]: E1124 06:55:35.851309 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:55:35 crc kubenswrapper[4809]: E1124 06:55:35.851465 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:51.851441531 +0000 UTC m=+51.752033236 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.890546 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.890614 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.890561 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:35 crc kubenswrapper[4809]: E1124 06:55:35.890723 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:55:35 crc kubenswrapper[4809]: E1124 06:55:35.890664 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:55:35 crc kubenswrapper[4809]: E1124 06:55:35.890771 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.930087 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.930135 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.930149 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.930166 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.930183 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:35Z","lastTransitionTime":"2025-11-24T06:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.951530 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:35 crc kubenswrapper[4809]: I1124 06:55:35.951575 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:35 crc kubenswrapper[4809]: E1124 06:55:35.951724 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:55:35 crc kubenswrapper[4809]: E1124 06:55:35.951743 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:55:35 crc kubenswrapper[4809]: E1124 06:55:35.951754 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:35 crc kubenswrapper[4809]: E1124 06:55:35.951786 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:55:35 crc kubenswrapper[4809]: E1124 06:55:35.951835 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:55:35 crc kubenswrapper[4809]: E1124 06:55:35.951852 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:35 crc kubenswrapper[4809]: E1124 06:55:35.951801 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:51.951788764 +0000 UTC m=+51.852380469 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:35 crc kubenswrapper[4809]: E1124 06:55:35.951984 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:51.951931778 +0000 UTC m=+51.852523473 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.031895 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.031938 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.031954 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.031984 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.031996 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:36Z","lastTransitionTime":"2025-11-24T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.133342 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.133387 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.133396 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.133408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.133418 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:36Z","lastTransitionTime":"2025-11-24T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.235692 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.235747 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.235758 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.235772 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.235781 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:36Z","lastTransitionTime":"2025-11-24T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.255442 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs\") pod \"network-metrics-daemon-krmz8\" (UID: \"35b8615e-faa5-47b4-8c22-609d5fda9590\") " pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:55:36 crc kubenswrapper[4809]: E1124 06:55:36.255602 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:55:36 crc kubenswrapper[4809]: E1124 06:55:36.255665 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs podName:35b8615e-faa5-47b4-8c22-609d5fda9590 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:38.255646393 +0000 UTC m=+38.156238108 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs") pod "network-metrics-daemon-krmz8" (UID: "35b8615e-faa5-47b4-8c22-609d5fda9590") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.338246 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.338317 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.338351 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.338378 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.338400 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:36Z","lastTransitionTime":"2025-11-24T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.440866 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.440909 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.440918 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.440933 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.440942 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:36Z","lastTransitionTime":"2025-11-24T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.543980 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.544016 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.544028 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.544042 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.544051 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:36Z","lastTransitionTime":"2025-11-24T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.646374 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.646435 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.646444 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.646460 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.646469 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:36Z","lastTransitionTime":"2025-11-24T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.749913 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.749948 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.749958 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.749996 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.750010 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:36Z","lastTransitionTime":"2025-11-24T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.798319 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.798401 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.798422 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.798448 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.798466 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:36Z","lastTransitionTime":"2025-11-24T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:36 crc kubenswrapper[4809]: E1124 06:55:36.816710 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:36Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.824333 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.824532 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.824645 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.824745 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.824835 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:36Z","lastTransitionTime":"2025-11-24T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:36 crc kubenswrapper[4809]: E1124 06:55:36.842781 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:36Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.847771 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.847804 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.847814 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.847844 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.847856 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:36Z","lastTransitionTime":"2025-11-24T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:36 crc kubenswrapper[4809]: E1124 06:55:36.865900 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:36Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.872637 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.872833 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.872873 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.872897 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.872917 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:36Z","lastTransitionTime":"2025-11-24T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:36 crc kubenswrapper[4809]: E1124 06:55:36.889496 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:36Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.890569 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:36 crc kubenswrapper[4809]: E1124 06:55:36.890776 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.894433 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.894504 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.894530 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.894560 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.894586 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:36Z","lastTransitionTime":"2025-11-24T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:36 crc kubenswrapper[4809]: E1124 06:55:36.911871 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:36Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:36 crc kubenswrapper[4809]: E1124 06:55:36.911999 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.914014 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.914045 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.914055 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.914075 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:36 crc kubenswrapper[4809]: I1124 06:55:36.914094 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:36Z","lastTransitionTime":"2025-11-24T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.017157 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.017250 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.017268 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.017291 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.017308 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:37Z","lastTransitionTime":"2025-11-24T06:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.120247 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.120307 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.120321 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.120345 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.120358 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:37Z","lastTransitionTime":"2025-11-24T06:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.223762 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.223810 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.223826 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.223847 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.223867 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:37Z","lastTransitionTime":"2025-11-24T06:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.326885 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.326924 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.326934 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.326949 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.326958 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:37Z","lastTransitionTime":"2025-11-24T06:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.429753 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.429807 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.429815 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.429829 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.429838 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:37Z","lastTransitionTime":"2025-11-24T06:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.532666 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.532727 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.532745 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.532769 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.532787 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:37Z","lastTransitionTime":"2025-11-24T06:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.635646 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.635722 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.635739 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.635768 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.635785 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:37Z","lastTransitionTime":"2025-11-24T06:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.740079 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.740130 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.740138 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.740154 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.740165 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:37Z","lastTransitionTime":"2025-11-24T06:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.844079 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.844133 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.844143 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.844163 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.844173 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:37Z","lastTransitionTime":"2025-11-24T06:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.890924 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.890941 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.891160 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:55:37 crc kubenswrapper[4809]: E1124 06:55:37.891271 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:55:37 crc kubenswrapper[4809]: E1124 06:55:37.891416 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:55:37 crc kubenswrapper[4809]: E1124 06:55:37.891666 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.946531 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.946593 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.946604 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.946619 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:37 crc kubenswrapper[4809]: I1124 06:55:37.946635 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:37Z","lastTransitionTime":"2025-11-24T06:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.049241 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.049284 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.049299 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.049315 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.049327 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:38Z","lastTransitionTime":"2025-11-24T06:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.152486 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.152552 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.152573 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.152599 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.152616 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:38Z","lastTransitionTime":"2025-11-24T06:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.255346 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.255396 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.255406 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.255421 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.255430 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:38Z","lastTransitionTime":"2025-11-24T06:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.275143 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs\") pod \"network-metrics-daemon-krmz8\" (UID: \"35b8615e-faa5-47b4-8c22-609d5fda9590\") " pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:55:38 crc kubenswrapper[4809]: E1124 06:55:38.275329 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:55:38 crc kubenswrapper[4809]: E1124 06:55:38.275402 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs podName:35b8615e-faa5-47b4-8c22-609d5fda9590 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:42.275385793 +0000 UTC m=+42.175977498 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs") pod "network-metrics-daemon-krmz8" (UID: "35b8615e-faa5-47b4-8c22-609d5fda9590") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.357783 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.357853 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.357865 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.357903 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.357913 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:38Z","lastTransitionTime":"2025-11-24T06:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.462446 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.462508 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.462532 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.462560 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.462583 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:38Z","lastTransitionTime":"2025-11-24T06:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.565610 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.565674 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.565697 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.565730 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.565752 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:38Z","lastTransitionTime":"2025-11-24T06:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.669785 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.669844 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.669861 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.669883 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.669900 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:38Z","lastTransitionTime":"2025-11-24T06:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.772402 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.772450 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.772468 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.772491 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.772507 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:38Z","lastTransitionTime":"2025-11-24T06:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.875131 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.875196 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.875218 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.875244 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.875265 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:38Z","lastTransitionTime":"2025-11-24T06:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.890826 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:38 crc kubenswrapper[4809]: E1124 06:55:38.891040 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.978406 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.978472 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.978491 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.978515 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:38 crc kubenswrapper[4809]: I1124 06:55:38.978534 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:38Z","lastTransitionTime":"2025-11-24T06:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.081233 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.081328 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.081346 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.081371 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.081389 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:39Z","lastTransitionTime":"2025-11-24T06:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.183776 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.183817 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.183826 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.183841 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.183852 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:39Z","lastTransitionTime":"2025-11-24T06:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.286707 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.286775 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.286799 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.286827 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.286850 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:39Z","lastTransitionTime":"2025-11-24T06:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.389146 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.389199 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.389213 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.389233 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.389249 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:39Z","lastTransitionTime":"2025-11-24T06:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.492211 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.492276 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.492293 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.492321 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.492339 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:39Z","lastTransitionTime":"2025-11-24T06:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.595472 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.595851 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.595881 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.595904 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.595921 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:39Z","lastTransitionTime":"2025-11-24T06:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.698281 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.698570 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.698664 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.698757 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.698850 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:39Z","lastTransitionTime":"2025-11-24T06:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.801477 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.801562 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.801579 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.801601 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.801617 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:39Z","lastTransitionTime":"2025-11-24T06:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.891109 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:39 crc kubenswrapper[4809]: E1124 06:55:39.891540 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.891169 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:39 crc kubenswrapper[4809]: E1124 06:55:39.891890 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.891124 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:55:39 crc kubenswrapper[4809]: E1124 06:55:39.892307 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.903864 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.903938 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.903993 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.904026 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:39 crc kubenswrapper[4809]: I1124 06:55:39.904049 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:39Z","lastTransitionTime":"2025-11-24T06:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.006674 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.006752 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.006771 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.006795 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.006814 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:40Z","lastTransitionTime":"2025-11-24T06:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.109566 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.109630 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.109653 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.109679 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.109696 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:40Z","lastTransitionTime":"2025-11-24T06:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.212191 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.212226 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.212234 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.212249 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.212259 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:40Z","lastTransitionTime":"2025-11-24T06:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.315564 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.315635 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.315660 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.315692 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.315714 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:40Z","lastTransitionTime":"2025-11-24T06:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.418823 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.418906 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.418926 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.418957 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.419012 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:40Z","lastTransitionTime":"2025-11-24T06:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.522322 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.522392 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.522408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.522436 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.522451 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:40Z","lastTransitionTime":"2025-11-24T06:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.625789 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.625872 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.625891 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.625914 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.625932 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:40Z","lastTransitionTime":"2025-11-24T06:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.729112 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.729179 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.729201 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.729226 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.729243 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:40Z","lastTransitionTime":"2025-11-24T06:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.832009 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.832079 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.832102 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.832131 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.832151 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:40Z","lastTransitionTime":"2025-11-24T06:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.891268 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:40 crc kubenswrapper[4809]: E1124 06:55:40.891752 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.920478 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:40Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.935615 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.935753 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.935822 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.935840 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.935853 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:40Z","lastTransitionTime":"2025-11-24T06:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.937616 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:40Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.955547 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:40Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:40 crc kubenswrapper[4809]: I1124 06:55:40.974752 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:40Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.001739 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:40Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.022276 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:41Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.038391 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.038454 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.038478 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.038443 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:41Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.038506 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.038531 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:41Z","lastTransitionTime":"2025-11-24T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.049041 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:41Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.061649 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krmz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b8615e-faa5-47b4-8c22-609d5fda9590\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krmz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:41Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.074830 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:41Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.087278 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:41Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.099142 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:41Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.115646 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:41Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.128120 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d459b25-158a-4a2f-bff7-b7c643450c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc5172a4b40f579c4ae113095afcdc882c281d31c0ddf2f59f0593c0a8e1837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69b51611cd90f67fb36b564db5890aff54f52a1884d71d297f335d707ec8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5fvqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:41Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.142066 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.142148 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.142188 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.142213 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.142230 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:41Z","lastTransitionTime":"2025-11-24T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.149521 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:32Z\\\",\\\"message\\\":\\\"rce:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8444, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1124 06:55:32.828572 6221 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z]\\\\nI1124 06:55:32.828645 6221 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manage\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-kpqhl_openshift-ovn-kubernetes(74622ad8-44b0-44a7-afbd-e19bb9635520)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:41Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.161503 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cq8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15c4f8f9-fd68-4636-88cb-d967cb3e202c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d9fa21c921ed12c61aa6a10a4cd4a5c575109a77a6b7099106b76414358d1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbdhx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cq8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:41Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.244826 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.244874 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.244888 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.244905 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.244918 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:41Z","lastTransitionTime":"2025-11-24T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.348642 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.348729 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.348753 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.348782 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.348801 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:41Z","lastTransitionTime":"2025-11-24T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.451010 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.451043 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.451052 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.451070 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.451082 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:41Z","lastTransitionTime":"2025-11-24T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.555912 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.556013 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.556031 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.556059 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.556070 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:41Z","lastTransitionTime":"2025-11-24T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.659364 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.659412 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.659428 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.659452 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.659470 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:41Z","lastTransitionTime":"2025-11-24T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.761597 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.761626 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.761635 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.761653 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.761665 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:41Z","lastTransitionTime":"2025-11-24T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.864324 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.864394 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.864406 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.864419 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.864428 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:41Z","lastTransitionTime":"2025-11-24T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.891229 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.891266 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:41 crc kubenswrapper[4809]: E1124 06:55:41.891412 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.891505 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:41 crc kubenswrapper[4809]: E1124 06:55:41.891793 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:55:41 crc kubenswrapper[4809]: E1124 06:55:41.891891 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.967387 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.967461 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.967477 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.967501 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:41 crc kubenswrapper[4809]: I1124 06:55:41.967514 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:41Z","lastTransitionTime":"2025-11-24T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.071129 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.071241 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.071884 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.072166 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.072232 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:42Z","lastTransitionTime":"2025-11-24T06:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.175331 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.175368 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.175377 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.175390 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.175399 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:42Z","lastTransitionTime":"2025-11-24T06:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.278203 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.278255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.278265 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.278282 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.278295 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:42Z","lastTransitionTime":"2025-11-24T06:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.333649 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs\") pod \"network-metrics-daemon-krmz8\" (UID: \"35b8615e-faa5-47b4-8c22-609d5fda9590\") " pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:55:42 crc kubenswrapper[4809]: E1124 06:55:42.333859 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:55:42 crc kubenswrapper[4809]: E1124 06:55:42.334003 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs podName:35b8615e-faa5-47b4-8c22-609d5fda9590 nodeName:}" failed. No retries permitted until 2025-11-24 06:55:50.333947132 +0000 UTC m=+50.234538927 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs") pod "network-metrics-daemon-krmz8" (UID: "35b8615e-faa5-47b4-8c22-609d5fda9590") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.381175 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.381208 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.381216 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.381230 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.381239 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:42Z","lastTransitionTime":"2025-11-24T06:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.483431 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.483467 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.483478 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.483492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.483502 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:42Z","lastTransitionTime":"2025-11-24T06:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.585839 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.585877 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.585886 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.585898 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.585908 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:42Z","lastTransitionTime":"2025-11-24T06:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.688663 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.688831 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.688853 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.688878 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.688894 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:42Z","lastTransitionTime":"2025-11-24T06:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.791910 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.792009 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.792038 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.792067 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.792088 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:42Z","lastTransitionTime":"2025-11-24T06:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.891032 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:42 crc kubenswrapper[4809]: E1124 06:55:42.891671 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.894085 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.894299 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.894473 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.894674 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.894838 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:42Z","lastTransitionTime":"2025-11-24T06:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.998179 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.998217 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.998229 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.998244 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:42 crc kubenswrapper[4809]: I1124 06:55:42.998255 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:42Z","lastTransitionTime":"2025-11-24T06:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.101211 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.101294 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.101319 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.101354 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.101375 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:43Z","lastTransitionTime":"2025-11-24T06:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.204699 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.204760 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.204777 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.204799 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.204822 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:43Z","lastTransitionTime":"2025-11-24T06:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.307844 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.307879 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.307889 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.307905 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.307914 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:43Z","lastTransitionTime":"2025-11-24T06:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.410953 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.411021 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.411034 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.411049 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.411060 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:43Z","lastTransitionTime":"2025-11-24T06:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.514624 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.514689 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.514711 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.514741 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.514764 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:43Z","lastTransitionTime":"2025-11-24T06:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.617662 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.617714 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.617726 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.617743 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.617756 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:43Z","lastTransitionTime":"2025-11-24T06:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.720383 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.720632 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.720716 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.720799 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.720879 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:43Z","lastTransitionTime":"2025-11-24T06:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.822801 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.823089 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.823188 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.823280 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.823360 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:43Z","lastTransitionTime":"2025-11-24T06:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.890399 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.890411 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.890458 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:43 crc kubenswrapper[4809]: E1124 06:55:43.890876 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:55:43 crc kubenswrapper[4809]: E1124 06:55:43.890725 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:55:43 crc kubenswrapper[4809]: E1124 06:55:43.891000 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.925819 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.925893 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.925907 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.925923 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:43 crc kubenswrapper[4809]: I1124 06:55:43.925936 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:43Z","lastTransitionTime":"2025-11-24T06:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.032454 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.032534 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.032559 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.032589 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.032612 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:44Z","lastTransitionTime":"2025-11-24T06:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.134814 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.134865 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.134880 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.134897 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.134909 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:44Z","lastTransitionTime":"2025-11-24T06:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.237596 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.237641 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.237650 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.237666 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.237677 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:44Z","lastTransitionTime":"2025-11-24T06:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.339891 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.339947 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.339957 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.339992 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.340005 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:44Z","lastTransitionTime":"2025-11-24T06:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.442360 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.442413 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.442426 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.442440 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.442451 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:44Z","lastTransitionTime":"2025-11-24T06:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.544872 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.544947 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.544985 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.545013 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.545026 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:44Z","lastTransitionTime":"2025-11-24T06:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.638604 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.640242 4809 scope.go:117] "RemoveContainer" containerID="743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.646634 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.646672 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.646682 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.646697 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.646707 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:44Z","lastTransitionTime":"2025-11-24T06:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.748913 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.748991 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.749008 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.749024 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.749034 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:44Z","lastTransitionTime":"2025-11-24T06:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.851285 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.851319 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.851327 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.851343 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.851353 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:44Z","lastTransitionTime":"2025-11-24T06:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.890933 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:44 crc kubenswrapper[4809]: E1124 06:55:44.891092 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.954205 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.954249 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.954263 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.954281 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:44 crc kubenswrapper[4809]: I1124 06:55:44.954295 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:44Z","lastTransitionTime":"2025-11-24T06:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.056888 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.056931 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.056945 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.056978 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.056991 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:45Z","lastTransitionTime":"2025-11-24T06:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.158580 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kpqhl_74622ad8-44b0-44a7-afbd-e19bb9635520/ovnkube-controller/1.log" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.158604 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.158636 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.158651 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.158671 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.158685 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:45Z","lastTransitionTime":"2025-11-24T06:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.160522 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerStarted","Data":"1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe"} Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.161584 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.174760 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.188494 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.199125 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.206950 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.214790 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.224623 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.246556 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.259448 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.261005 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.261053 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.261067 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.261086 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.261097 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:45Z","lastTransitionTime":"2025-11-24T06:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.284380 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krmz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b8615e-faa5-47b4-8c22-609d5fda9590\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krmz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.295507 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.306539 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.317419 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.327622 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.337344 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d459b25-158a-4a2f-bff7-b7c643450c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc5172a4b40f579c4ae113095afcdc882c281d31c0ddf2f59f0593c0a8e1837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69b51611cd90f67fb36b564db5890aff54f52a1884d71d297f335d707ec8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5fvqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.353363 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:32Z\\\",\\\"message\\\":\\\"rce:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8444, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1124 06:55:32.828572 6221 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z]\\\\nI1124 06:55:32.828645 6221 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manage\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.361998 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cq8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15c4f8f9-fd68-4636-88cb-d967cb3e202c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d9fa21c921ed12c61aa6a10a4cd4a5c575109a77a6b7099106b76414358d1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbdhx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cq8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.363551 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.363578 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.363590 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.363608 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.363627 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:45Z","lastTransitionTime":"2025-11-24T06:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.466626 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.466666 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.466682 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.466884 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.466907 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:45Z","lastTransitionTime":"2025-11-24T06:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.569868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.569904 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.569921 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.569938 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.569950 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:45Z","lastTransitionTime":"2025-11-24T06:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.674944 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.675237 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.675262 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.675297 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.675552 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:45Z","lastTransitionTime":"2025-11-24T06:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.779041 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.779079 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.779092 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.779107 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.779314 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:45Z","lastTransitionTime":"2025-11-24T06:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.881691 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.881720 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.881728 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.881753 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.881762 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:45Z","lastTransitionTime":"2025-11-24T06:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.890474 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.890601 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:45 crc kubenswrapper[4809]: E1124 06:55:45.890663 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:55:45 crc kubenswrapper[4809]: E1124 06:55:45.890741 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:55:45 crc kubenswrapper[4809]: I1124 06:55:45.890721 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:55:46 crc kubenswrapper[4809]: E1124 06:55:46.942915 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:55:46 crc kubenswrapper[4809]: I1124 06:55:46.943981 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:46 crc kubenswrapper[4809]: E1124 06:55:46.944098 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:55:46 crc kubenswrapper[4809]: I1124 06:55:46.949158 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:46 crc kubenswrapper[4809]: I1124 06:55:46.949214 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:46 crc kubenswrapper[4809]: I1124 06:55:46.949232 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:46 crc kubenswrapper[4809]: I1124 06:55:46.949259 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:46 crc kubenswrapper[4809]: I1124 06:55:46.949278 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:46Z","lastTransitionTime":"2025-11-24T06:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.052321 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.052363 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.052372 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.052409 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.052424 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:47Z","lastTransitionTime":"2025-11-24T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.154921 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.155053 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.155084 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.155119 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.155140 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:47Z","lastTransitionTime":"2025-11-24T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.258306 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.258339 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.258346 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.258359 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.258369 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:47Z","lastTransitionTime":"2025-11-24T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.313601 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.313648 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.313656 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.313670 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.313683 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:47Z","lastTransitionTime":"2025-11-24T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:47 crc kubenswrapper[4809]: E1124 06:55:47.332433 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:47Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.336670 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.336731 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.336742 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.336757 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.336766 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:47Z","lastTransitionTime":"2025-11-24T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:47 crc kubenswrapper[4809]: E1124 06:55:47.350578 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:47Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.355194 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.355260 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.355273 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.355290 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.355302 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:47Z","lastTransitionTime":"2025-11-24T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:47 crc kubenswrapper[4809]: E1124 06:55:47.368186 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:47Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.372297 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.372345 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.372361 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.372379 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.372391 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:47Z","lastTransitionTime":"2025-11-24T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:47 crc kubenswrapper[4809]: E1124 06:55:47.403842 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:47Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.413038 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.413081 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.413093 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.413111 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.413126 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:47Z","lastTransitionTime":"2025-11-24T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:47 crc kubenswrapper[4809]: E1124 06:55:47.432334 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:47Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:47 crc kubenswrapper[4809]: E1124 06:55:47.432584 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.434288 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.434324 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.434338 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.434355 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.434368 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:47Z","lastTransitionTime":"2025-11-24T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.536698 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.536767 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.536785 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.536809 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.537066 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:47Z","lastTransitionTime":"2025-11-24T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.639808 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.639863 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.639880 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.639902 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.639918 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:47Z","lastTransitionTime":"2025-11-24T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.742347 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.742414 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.742435 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.742465 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.742487 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:47Z","lastTransitionTime":"2025-11-24T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.845710 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.845768 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.845786 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.845809 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.845826 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:47Z","lastTransitionTime":"2025-11-24T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.890408 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.890472 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:47 crc kubenswrapper[4809]: E1124 06:55:47.890614 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:55:47 crc kubenswrapper[4809]: E1124 06:55:47.890737 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.949473 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.949530 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.949550 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.949628 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.949648 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:47Z","lastTransitionTime":"2025-11-24T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.954599 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kpqhl_74622ad8-44b0-44a7-afbd-e19bb9635520/ovnkube-controller/2.log" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.955515 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kpqhl_74622ad8-44b0-44a7-afbd-e19bb9635520/ovnkube-controller/1.log" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.958935 4809 generic.go:334] "Generic (PLEG): container finished" podID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerID="1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe" exitCode=1 Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.959000 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerDied","Data":"1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe"} Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.959051 4809 scope.go:117] "RemoveContainer" containerID="743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.960306 4809 scope.go:117] "RemoveContainer" containerID="1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe" Nov 24 06:55:47 crc kubenswrapper[4809]: E1124 06:55:47.960581 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kpqhl_openshift-ovn-kubernetes(74622ad8-44b0-44a7-afbd-e19bb9635520)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.980166 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:47Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:47 crc kubenswrapper[4809]: I1124 06:55:47.998832 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:47Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.016427 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:48Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.033258 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:48Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.048801 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:48Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.052847 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.052914 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.052925 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.052939 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.052948 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:48Z","lastTransitionTime":"2025-11-24T06:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.064130 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:48Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.086221 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:48Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.105907 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:48Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.118811 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krmz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b8615e-faa5-47b4-8c22-609d5fda9590\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krmz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:48Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.134428 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:48Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.149296 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:48Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.155684 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.155729 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.155742 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.155757 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.156104 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:48Z","lastTransitionTime":"2025-11-24T06:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.161039 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:48Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.181054 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:48Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.194898 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d459b25-158a-4a2f-bff7-b7c643450c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc5172a4b40f579c4ae113095afcdc882c281d31c0ddf2f59f0593c0a8e1837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69b51611cd90f67fb36b564db5890aff54f52a1884d71d297f335d707ec8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5fvqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:48Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.216760 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:32Z\\\",\\\"message\\\":\\\"rce:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8444, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1124 06:55:32.828572 6221 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z]\\\\nI1124 06:55:32.828645 6221 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manage\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"lector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127759 6475 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127828 6475 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127886 6475 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127949 6475 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.128093 6475 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.128496 6475 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 06:55:47.128522 6475 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 06:55:47.128544 6475 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 06:55:47.128555 6475 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 06:55:47.128583 6475 factory.go:656] Stopping watch factory\\\\nI1124 06:55:47.128600 6475 ovnkube.go:599] Stopped ovnkube\\\\nI1124 06:55:47.128621 6475 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:48Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.227164 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cq8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15c4f8f9-fd68-4636-88cb-d967cb3e202c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d9fa21c921ed12c61aa6a10a4cd4a5c575109a77a6b7099106b76414358d1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbdhx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cq8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:48Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.257926 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.257997 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.258012 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.258029 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.258042 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:48Z","lastTransitionTime":"2025-11-24T06:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.361032 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.361089 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.361106 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.361126 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.361137 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:48Z","lastTransitionTime":"2025-11-24T06:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.463940 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.464012 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.464024 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.464041 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.464050 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:48Z","lastTransitionTime":"2025-11-24T06:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.566493 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.566526 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.566537 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.566549 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.566559 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:48Z","lastTransitionTime":"2025-11-24T06:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.669123 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.669192 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.669214 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.669247 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.669271 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:48Z","lastTransitionTime":"2025-11-24T06:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.772391 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.772443 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.772464 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.772486 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.772504 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:48Z","lastTransitionTime":"2025-11-24T06:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.875945 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.876002 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.876011 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.876027 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.876038 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:48Z","lastTransitionTime":"2025-11-24T06:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.890488 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.890591 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:55:48 crc kubenswrapper[4809]: E1124 06:55:48.890719 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:55:48 crc kubenswrapper[4809]: E1124 06:55:48.890800 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.963399 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kpqhl_74622ad8-44b0-44a7-afbd-e19bb9635520/ovnkube-controller/2.log" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.977851 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.977881 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.978098 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.978130 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:48 crc kubenswrapper[4809]: I1124 06:55:48.978141 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:48Z","lastTransitionTime":"2025-11-24T06:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.080253 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.080290 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.080298 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.080326 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.080337 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:49Z","lastTransitionTime":"2025-11-24T06:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.183203 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.183245 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.183258 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.183276 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.183288 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:49Z","lastTransitionTime":"2025-11-24T06:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.285021 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.285060 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.285068 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.285086 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.285096 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:49Z","lastTransitionTime":"2025-11-24T06:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.387704 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.387783 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.387809 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.387922 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.387949 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:49Z","lastTransitionTime":"2025-11-24T06:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.490060 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.490111 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.490129 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.490147 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.490160 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:49Z","lastTransitionTime":"2025-11-24T06:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.593249 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.593290 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.593300 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.593313 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.593322 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:49Z","lastTransitionTime":"2025-11-24T06:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.697060 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.697120 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.697162 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.697201 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.697231 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:49Z","lastTransitionTime":"2025-11-24T06:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.800604 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.800675 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.800698 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.800727 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.800750 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:49Z","lastTransitionTime":"2025-11-24T06:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.890544 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.890540 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:49 crc kubenswrapper[4809]: E1124 06:55:49.890763 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:55:49 crc kubenswrapper[4809]: E1124 06:55:49.890911 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.904215 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.904272 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.904294 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.904321 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:49 crc kubenswrapper[4809]: I1124 06:55:49.904343 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:49Z","lastTransitionTime":"2025-11-24T06:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.007478 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.007575 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.007594 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.007615 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.007632 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:50Z","lastTransitionTime":"2025-11-24T06:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.111926 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.111988 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.112001 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.112017 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.112029 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:50Z","lastTransitionTime":"2025-11-24T06:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.215383 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.215430 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.215447 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.215471 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.215492 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:50Z","lastTransitionTime":"2025-11-24T06:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.318360 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.318405 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.318421 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.318442 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.318460 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:50Z","lastTransitionTime":"2025-11-24T06:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.374009 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs\") pod \"network-metrics-daemon-krmz8\" (UID: \"35b8615e-faa5-47b4-8c22-609d5fda9590\") " pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:55:50 crc kubenswrapper[4809]: E1124 06:55:50.374278 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:55:50 crc kubenswrapper[4809]: E1124 06:55:50.374421 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs podName:35b8615e-faa5-47b4-8c22-609d5fda9590 nodeName:}" failed. No retries permitted until 2025-11-24 06:56:06.374384451 +0000 UTC m=+66.274976226 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs") pod "network-metrics-daemon-krmz8" (UID: "35b8615e-faa5-47b4-8c22-609d5fda9590") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.420834 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.420865 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.420873 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.420886 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.420894 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:50Z","lastTransitionTime":"2025-11-24T06:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.523644 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.523688 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.523702 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.523723 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.523745 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:50Z","lastTransitionTime":"2025-11-24T06:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.626750 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.626788 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.626799 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.626818 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.626833 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:50Z","lastTransitionTime":"2025-11-24T06:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.729783 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.729850 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.729869 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.729894 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.729911 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:50Z","lastTransitionTime":"2025-11-24T06:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.832524 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.832588 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.832610 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.832636 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.832655 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:50Z","lastTransitionTime":"2025-11-24T06:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.890776 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.891116 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:50 crc kubenswrapper[4809]: E1124 06:55:50.891322 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:55:50 crc kubenswrapper[4809]: E1124 06:55:50.891625 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.910470 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:50Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.932295 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:50Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.934855 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.934907 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.934919 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.934936 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.934949 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:50Z","lastTransitionTime":"2025-11-24T06:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.953082 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:50Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.972096 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:50Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:50 crc kubenswrapper[4809]: I1124 06:55:50.988611 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d459b25-158a-4a2f-bff7-b7c643450c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc5172a4b40f579c4ae113095afcdc882c281d31c0ddf2f59f0593c0a8e1837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69b51611cd90f67fb36b564db5890aff54f52a1884d71d297f335d707ec8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5fvqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:50Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.013545 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:32Z\\\",\\\"message\\\":\\\"rce:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8444, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1124 06:55:32.828572 6221 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z]\\\\nI1124 06:55:32.828645 6221 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manage\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"lector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127759 6475 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127828 6475 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127886 6475 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127949 6475 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.128093 6475 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.128496 6475 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 06:55:47.128522 6475 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 06:55:47.128544 6475 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 06:55:47.128555 6475 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 06:55:47.128583 6475 factory.go:656] Stopping watch factory\\\\nI1124 06:55:47.128600 6475 ovnkube.go:599] Stopped ovnkube\\\\nI1124 06:55:47.128621 6475 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:51Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.027092 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cq8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15c4f8f9-fd68-4636-88cb-d967cb3e202c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d9fa21c921ed12c61aa6a10a4cd4a5c575109a77a6b7099106b76414358d1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbdhx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cq8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:51Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.037471 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.037550 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.037565 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.037585 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.037599 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:51Z","lastTransitionTime":"2025-11-24T06:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.040674 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:51Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.050842 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:51Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.066776 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:51Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.083764 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:51Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.096911 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:51Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.108840 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:51Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.123074 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:51Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.137075 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:51Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.139741 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.139781 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.139793 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.139833 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.139845 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:51Z","lastTransitionTime":"2025-11-24T06:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.148242 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krmz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b8615e-faa5-47b4-8c22-609d5fda9590\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krmz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:51Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.242246 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.242300 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.242315 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.242336 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.242352 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:51Z","lastTransitionTime":"2025-11-24T06:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.345179 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.345215 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.345223 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.345237 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.345248 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:51Z","lastTransitionTime":"2025-11-24T06:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.448068 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.448145 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.448166 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.448195 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.448217 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:51Z","lastTransitionTime":"2025-11-24T06:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.551569 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.551625 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.551633 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.551649 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.551660 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:51Z","lastTransitionTime":"2025-11-24T06:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.654667 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.654726 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.654739 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.654756 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.654769 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:51Z","lastTransitionTime":"2025-11-24T06:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.758034 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.758113 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.758141 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.758169 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.758187 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:51Z","lastTransitionTime":"2025-11-24T06:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.860390 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.860434 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.860442 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.860458 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.860467 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:51Z","lastTransitionTime":"2025-11-24T06:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.889631 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:55:51 crc kubenswrapper[4809]: E1124 06:55:51.889817 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:56:23.889781151 +0000 UTC m=+83.790372906 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.889893 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.890030 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:51 crc kubenswrapper[4809]: E1124 06:55:51.890155 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:55:51 crc kubenswrapper[4809]: E1124 06:55:51.890246 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:56:23.890221783 +0000 UTC m=+83.790813528 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:55:51 crc kubenswrapper[4809]: E1124 06:55:51.890272 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:55:51 crc kubenswrapper[4809]: E1124 06:55:51.890371 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:56:23.890342077 +0000 UTC m=+83.790933822 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.890436 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.890454 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:51 crc kubenswrapper[4809]: E1124 06:55:51.890619 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:55:51 crc kubenswrapper[4809]: E1124 06:55:51.890695 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.963186 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.963220 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.963227 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.963239 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.963248 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:51Z","lastTransitionTime":"2025-11-24T06:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.990559 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:51 crc kubenswrapper[4809]: I1124 06:55:51.990606 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:51 crc kubenswrapper[4809]: E1124 06:55:51.990742 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:55:51 crc kubenswrapper[4809]: E1124 06:55:51.990764 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:55:51 crc kubenswrapper[4809]: E1124 06:55:51.990777 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:51 crc kubenswrapper[4809]: E1124 06:55:51.990818 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 06:56:23.990803831 +0000 UTC m=+83.891395536 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:51 crc kubenswrapper[4809]: E1124 06:55:51.990742 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:55:51 crc kubenswrapper[4809]: E1124 06:55:51.990864 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:55:51 crc kubenswrapper[4809]: E1124 06:55:51.990878 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:51 crc kubenswrapper[4809]: E1124 06:55:51.990911 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 06:56:23.990900373 +0000 UTC m=+83.891492078 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.065582 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.065621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.065630 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.065643 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.065652 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:52Z","lastTransitionTime":"2025-11-24T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.168496 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.168534 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.168545 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.168560 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.168573 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:52Z","lastTransitionTime":"2025-11-24T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.272405 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.272457 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.272471 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.272489 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.272501 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:52Z","lastTransitionTime":"2025-11-24T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.374870 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.374918 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.374934 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.374960 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.374989 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:52Z","lastTransitionTime":"2025-11-24T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.478131 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.478457 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.478469 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.478487 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.478499 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:52Z","lastTransitionTime":"2025-11-24T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.581604 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.581656 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.581672 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.581695 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.581713 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:52Z","lastTransitionTime":"2025-11-24T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.685237 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.685306 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.685321 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.685348 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.685368 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:52Z","lastTransitionTime":"2025-11-24T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.788462 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.788581 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.788603 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.788630 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.788648 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:52Z","lastTransitionTime":"2025-11-24T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.890172 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.890239 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:52 crc kubenswrapper[4809]: E1124 06:55:52.890347 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:55:52 crc kubenswrapper[4809]: E1124 06:55:52.890536 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.891618 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.891723 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.891743 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.891809 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.891829 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:52Z","lastTransitionTime":"2025-11-24T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.994794 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.994879 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.994911 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.994940 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:52 crc kubenswrapper[4809]: I1124 06:55:52.995010 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:52Z","lastTransitionTime":"2025-11-24T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.097470 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.097533 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.097551 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.097577 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.097597 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:53Z","lastTransitionTime":"2025-11-24T06:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.200419 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.200557 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.200569 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.200584 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.200595 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:53Z","lastTransitionTime":"2025-11-24T06:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.303589 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.303666 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.303685 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.303712 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.303732 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:53Z","lastTransitionTime":"2025-11-24T06:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.407692 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.407759 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.407779 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.407805 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.407824 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:53Z","lastTransitionTime":"2025-11-24T06:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.511547 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.511607 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.511628 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.511655 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.511676 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:53Z","lastTransitionTime":"2025-11-24T06:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.615297 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.615377 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.615401 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.615429 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.615447 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:53Z","lastTransitionTime":"2025-11-24T06:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.717647 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.717708 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.717725 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.717748 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.717765 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:53Z","lastTransitionTime":"2025-11-24T06:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.820791 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.820865 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.820887 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.820912 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.820929 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:53Z","lastTransitionTime":"2025-11-24T06:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.890636 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.890696 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:53 crc kubenswrapper[4809]: E1124 06:55:53.890790 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:55:53 crc kubenswrapper[4809]: E1124 06:55:53.890930 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.924296 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.924371 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.924396 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.924427 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:53 crc kubenswrapper[4809]: I1124 06:55:53.924480 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:53Z","lastTransitionTime":"2025-11-24T06:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.027673 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.027715 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.027728 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.027747 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.027761 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:54Z","lastTransitionTime":"2025-11-24T06:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.130882 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.130921 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.130931 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.130946 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.130957 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:54Z","lastTransitionTime":"2025-11-24T06:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.234371 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.234430 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.234447 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.234472 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.234489 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:54Z","lastTransitionTime":"2025-11-24T06:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.336730 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.336795 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.336816 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.336839 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.336856 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:54Z","lastTransitionTime":"2025-11-24T06:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.440109 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.440264 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.440284 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.440314 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.440351 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:54Z","lastTransitionTime":"2025-11-24T06:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.542567 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.542615 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.542630 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.542650 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.542664 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:54Z","lastTransitionTime":"2025-11-24T06:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.645938 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.646013 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.646026 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.646045 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.646058 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:54Z","lastTransitionTime":"2025-11-24T06:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.749792 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.749865 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.749893 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.749924 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.749950 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:54Z","lastTransitionTime":"2025-11-24T06:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.853609 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.853698 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.853729 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.853760 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.853784 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:54Z","lastTransitionTime":"2025-11-24T06:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.891227 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.891307 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:54 crc kubenswrapper[4809]: E1124 06:55:54.891483 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:55:54 crc kubenswrapper[4809]: E1124 06:55:54.891608 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.956257 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.956299 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.956311 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.956327 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:54 crc kubenswrapper[4809]: I1124 06:55:54.956339 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:54Z","lastTransitionTime":"2025-11-24T06:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.059030 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.059068 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.059077 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.059092 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.059103 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:55Z","lastTransitionTime":"2025-11-24T06:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.160999 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.161036 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.161048 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.161064 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.161075 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:55Z","lastTransitionTime":"2025-11-24T06:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.265034 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.265079 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.265099 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.265119 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.265130 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:55Z","lastTransitionTime":"2025-11-24T06:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.367647 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.367709 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.367726 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.367748 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.367769 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:55Z","lastTransitionTime":"2025-11-24T06:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.470536 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.470609 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.470627 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.470652 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.470669 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:55Z","lastTransitionTime":"2025-11-24T06:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.573707 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.573751 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.573760 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.573774 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.573783 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:55Z","lastTransitionTime":"2025-11-24T06:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.677314 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.677368 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.677402 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.677423 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.677437 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:55Z","lastTransitionTime":"2025-11-24T06:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.781029 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.781080 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.781093 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.781110 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.781121 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:55Z","lastTransitionTime":"2025-11-24T06:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.883460 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.883517 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.883535 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.883560 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.883580 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:55Z","lastTransitionTime":"2025-11-24T06:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.891059 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.891136 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:55 crc kubenswrapper[4809]: E1124 06:55:55.891207 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:55:55 crc kubenswrapper[4809]: E1124 06:55:55.891288 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.985610 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.985646 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.985657 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.985676 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:55 crc kubenswrapper[4809]: I1124 06:55:55.985689 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:55Z","lastTransitionTime":"2025-11-24T06:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.088218 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.088261 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.088272 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.088290 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.088301 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:56Z","lastTransitionTime":"2025-11-24T06:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.190645 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.190685 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.190708 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.190722 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.190731 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:56Z","lastTransitionTime":"2025-11-24T06:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.293296 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.293336 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.293347 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.293360 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.293369 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:56Z","lastTransitionTime":"2025-11-24T06:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.395929 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.396020 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.396043 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.396073 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.396097 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:56Z","lastTransitionTime":"2025-11-24T06:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.498528 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.498596 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.498614 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.498638 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.498658 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:56Z","lastTransitionTime":"2025-11-24T06:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.601335 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.601399 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.601420 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.601447 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.601469 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:56Z","lastTransitionTime":"2025-11-24T06:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.704204 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.704249 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.704265 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.704287 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.704304 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:56Z","lastTransitionTime":"2025-11-24T06:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.811182 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.811249 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.811261 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.811279 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.811295 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:56Z","lastTransitionTime":"2025-11-24T06:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.890639 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:55:56 crc kubenswrapper[4809]: E1124 06:55:56.890774 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.890648 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:56 crc kubenswrapper[4809]: E1124 06:55:56.890997 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.913671 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.913703 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.913713 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.913727 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:56 crc kubenswrapper[4809]: I1124 06:55:56.913737 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:56Z","lastTransitionTime":"2025-11-24T06:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.015742 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.015799 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.015812 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.015830 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.015842 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:57Z","lastTransitionTime":"2025-11-24T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.119482 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.119537 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.119548 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.119566 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.119577 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:57Z","lastTransitionTime":"2025-11-24T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.222911 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.222951 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.222959 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.222988 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.222998 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:57Z","lastTransitionTime":"2025-11-24T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.325303 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.325356 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.325369 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.325388 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.325401 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:57Z","lastTransitionTime":"2025-11-24T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.427902 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.427961 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.428014 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.428036 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.428053 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:57Z","lastTransitionTime":"2025-11-24T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.531062 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.531145 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.531169 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.531207 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.531232 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:57Z","lastTransitionTime":"2025-11-24T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.623550 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.634466 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.634515 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.634539 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.634571 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.634594 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:57Z","lastTransitionTime":"2025-11-24T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.642104 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.642242 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.663816 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.682118 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.697020 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.720361 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d459b25-158a-4a2f-bff7-b7c643450c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc5172a4b40f579c4ae113095afcdc882c281d31c0ddf2f59f0593c0a8e1837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69b51611cd90f67fb36b564db5890aff54f52a1884d71d297f335d707ec8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5fvqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.737761 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cq8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15c4f8f9-fd68-4636-88cb-d967cb3e202c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d9fa21c921ed12c61aa6a10a4cd4a5c575109a77a6b7099106b76414358d1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbdhx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cq8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.739754 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.739841 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.739863 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.739908 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.740040 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:57Z","lastTransitionTime":"2025-11-24T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.763854 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.763898 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.763910 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.763926 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.763679 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:32Z\\\",\\\"message\\\":\\\"rce:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8444, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1124 06:55:32.828572 6221 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z]\\\\nI1124 06:55:32.828645 6221 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manage\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"lector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127759 6475 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127828 6475 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127886 6475 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127949 6475 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.128093 6475 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.128496 6475 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 06:55:47.128522 6475 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 06:55:47.128544 6475 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 06:55:47.128555 6475 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 06:55:47.128583 6475 factory.go:656] Stopping watch factory\\\\nI1124 06:55:47.128600 6475 ovnkube.go:599] Stopped ovnkube\\\\nI1124 06:55:47.128621 6475 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.763938 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:57Z","lastTransitionTime":"2025-11-24T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.777789 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:57 crc kubenswrapper[4809]: E1124 06:55:57.779045 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.783593 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.783647 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.783662 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.783686 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.783706 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:57Z","lastTransitionTime":"2025-11-24T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.791016 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:57 crc kubenswrapper[4809]: E1124 06:55:57.800722 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.802519 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.804531 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.804557 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.804567 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.804583 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.804595 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:57Z","lastTransitionTime":"2025-11-24T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.813546 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:57 crc kubenswrapper[4809]: E1124 06:55:57.821561 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.825343 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.825378 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.825388 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.825406 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.825419 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:57Z","lastTransitionTime":"2025-11-24T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.825462 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:57 crc kubenswrapper[4809]: E1124 06:55:57.838618 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.842336 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.842374 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.842401 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.842427 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.842439 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:57Z","lastTransitionTime":"2025-11-24T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.849430 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:57 crc kubenswrapper[4809]: E1124 06:55:57.861939 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:57 crc kubenswrapper[4809]: E1124 06:55:57.862087 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.862852 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.864046 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.864078 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.864089 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.864107 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.864119 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:57Z","lastTransitionTime":"2025-11-24T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.879184 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.890896 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.890955 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:57 crc kubenswrapper[4809]: E1124 06:55:57.891097 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:55:57 crc kubenswrapper[4809]: E1124 06:55:57.891220 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.892542 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krmz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b8615e-faa5-47b4-8c22-609d5fda9590\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krmz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.967129 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.967178 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.967188 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.967207 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:57 crc kubenswrapper[4809]: I1124 06:55:57.967217 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:57Z","lastTransitionTime":"2025-11-24T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.070461 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.070524 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.070539 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.070559 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.070573 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:58Z","lastTransitionTime":"2025-11-24T06:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.173630 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.173678 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.173689 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.173706 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.173719 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:58Z","lastTransitionTime":"2025-11-24T06:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.277094 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.277183 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.277209 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.277258 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.277285 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:58Z","lastTransitionTime":"2025-11-24T06:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.379918 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.379987 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.379997 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.380014 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.380024 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:58Z","lastTransitionTime":"2025-11-24T06:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.482727 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.482798 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.482819 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.482849 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.482875 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:58Z","lastTransitionTime":"2025-11-24T06:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.585960 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.586019 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.586029 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.586053 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.586076 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:58Z","lastTransitionTime":"2025-11-24T06:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.689283 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.689336 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.689349 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.689368 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.689381 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:58Z","lastTransitionTime":"2025-11-24T06:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.792375 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.792424 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.792437 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.792458 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.792472 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:58Z","lastTransitionTime":"2025-11-24T06:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.890910 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.891007 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:55:58 crc kubenswrapper[4809]: E1124 06:55:58.891267 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:55:58 crc kubenswrapper[4809]: E1124 06:55:58.891462 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.897394 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.897445 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.897466 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.897493 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:58 crc kubenswrapper[4809]: I1124 06:55:58.897514 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:58Z","lastTransitionTime":"2025-11-24T06:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.000934 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.001174 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.001198 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.001229 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.001247 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:59Z","lastTransitionTime":"2025-11-24T06:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.103832 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.103863 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.103872 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.103885 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.103895 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:59Z","lastTransitionTime":"2025-11-24T06:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.207102 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.207145 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.207172 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.207194 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.207209 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:59Z","lastTransitionTime":"2025-11-24T06:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.310573 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.310630 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.310654 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.310686 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.310710 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:59Z","lastTransitionTime":"2025-11-24T06:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.414439 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.414483 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.414503 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.414525 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.414541 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:59Z","lastTransitionTime":"2025-11-24T06:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.517853 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.517924 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.517945 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.518006 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.518030 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:59Z","lastTransitionTime":"2025-11-24T06:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.621034 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.621084 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.621097 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.621115 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.621128 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:59Z","lastTransitionTime":"2025-11-24T06:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.723525 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.723643 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.723671 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.723713 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.723745 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:59Z","lastTransitionTime":"2025-11-24T06:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.825806 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.825842 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.825850 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.825880 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.825889 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:59Z","lastTransitionTime":"2025-11-24T06:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.890903 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.890945 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:55:59 crc kubenswrapper[4809]: E1124 06:55:59.891081 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:55:59 crc kubenswrapper[4809]: E1124 06:55:59.891217 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.928264 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.928298 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.928308 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.928324 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:55:59 crc kubenswrapper[4809]: I1124 06:55:59.928335 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:55:59Z","lastTransitionTime":"2025-11-24T06:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.030990 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.031030 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.031041 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.031055 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.031066 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:00Z","lastTransitionTime":"2025-11-24T06:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.134079 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.134130 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.134141 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.134159 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.134171 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:00Z","lastTransitionTime":"2025-11-24T06:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.237119 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.237177 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.237188 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.237205 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.237217 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:00Z","lastTransitionTime":"2025-11-24T06:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.339433 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.339465 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.339476 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.339515 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.339528 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:00Z","lastTransitionTime":"2025-11-24T06:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.442142 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.442191 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.442202 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.442222 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.442239 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:00Z","lastTransitionTime":"2025-11-24T06:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.545241 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.545280 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.545288 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.545303 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.545312 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:00Z","lastTransitionTime":"2025-11-24T06:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.647782 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.647831 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.647842 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.647856 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.647868 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:00Z","lastTransitionTime":"2025-11-24T06:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.750567 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.750630 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.750652 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.750681 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.750702 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:00Z","lastTransitionTime":"2025-11-24T06:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.853288 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.853347 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.853365 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.853389 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.853408 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:00Z","lastTransitionTime":"2025-11-24T06:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.890376 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.890401 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:00 crc kubenswrapper[4809]: E1124 06:56:00.890566 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:00 crc kubenswrapper[4809]: E1124 06:56:00.890913 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.892167 4809 scope.go:117] "RemoveContainer" containerID="1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe" Nov 24 06:56:00 crc kubenswrapper[4809]: E1124 06:56:00.892518 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kpqhl_openshift-ovn-kubernetes(74622ad8-44b0-44a7-afbd-e19bb9635520)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.913062 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4e7ae8a-acd5-4685-97b4-9f775d5c08d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80b7097545fca095740245eb7373a8ad9325bfde309725cf0b0db792bbd44df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5376dc11940017714443d1a4cf6c61f323eb35b41d1ce9b76d73137465f18c83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70c7310e5734c08f62e5732d069aed4eb9ccd5cc0e8ae546950871772f6a27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66bb948c15a1089cc29548df9da9f67b1f554d907a02f6076a0da38cabca9670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66bb948c15a1089cc29548df9da9f67b1f554d907a02f6076a0da38cabca9670\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:00Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.928953 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:00Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.942635 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:00Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.954663 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:00Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.955575 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.955618 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.955632 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.955653 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.955668 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:00Z","lastTransitionTime":"2025-11-24T06:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.970328 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:00Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:00 crc kubenswrapper[4809]: I1124 06:56:00.987487 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d459b25-158a-4a2f-bff7-b7c643450c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc5172a4b40f579c4ae113095afcdc882c281d31c0ddf2f59f0593c0a8e1837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69b51611cd90f67fb36b564db5890aff54f52a1884d71d297f335d707ec8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5fvqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:00Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.010301 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://743dd415d6655f90719a53cf43616d0289f5e09cb46d59fab97141126cd1ed40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:32Z\\\",\\\"message\\\":\\\"rce:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8444, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1124 06:55:32.828572 6221 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:55:32Z is after 2025-08-24T17:21:41Z]\\\\nI1124 06:55:32.828645 6221 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manage\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"lector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127759 6475 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127828 6475 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127886 6475 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127949 6475 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.128093 6475 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.128496 6475 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 06:55:47.128522 6475 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 06:55:47.128544 6475 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 06:55:47.128555 6475 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 06:55:47.128583 6475 factory.go:656] Stopping watch factory\\\\nI1124 06:55:47.128600 6475 ovnkube.go:599] Stopped ovnkube\\\\nI1124 06:55:47.128621 6475 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.022234 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cq8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15c4f8f9-fd68-4636-88cb-d967cb3e202c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d9fa21c921ed12c61aa6a10a4cd4a5c575109a77a6b7099106b76414358d1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbdhx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cq8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.033997 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.047828 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.058652 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.058691 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.058700 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.058732 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.058742 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:01Z","lastTransitionTime":"2025-11-24T06:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.063441 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.078369 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.096443 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.110374 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.122348 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.135837 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.147551 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krmz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b8615e-faa5-47b4-8c22-609d5fda9590\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krmz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.161386 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.161453 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.161463 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.161477 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.161487 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:01Z","lastTransitionTime":"2025-11-24T06:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.163174 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.179824 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.196147 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.212997 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.224317 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.235368 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.248056 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.262117 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.264887 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.265029 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.265098 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.265167 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.265228 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:01Z","lastTransitionTime":"2025-11-24T06:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.274504 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krmz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b8615e-faa5-47b4-8c22-609d5fda9590\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krmz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.288507 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.303612 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.318316 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.331603 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.344856 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4e7ae8a-acd5-4685-97b4-9f775d5c08d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80b7097545fca095740245eb7373a8ad9325bfde309725cf0b0db792bbd44df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5376dc11940017714443d1a4cf6c61f323eb35b41d1ce9b76d73137465f18c83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70c7310e5734c08f62e5732d069aed4eb9ccd5cc0e8ae546950871772f6a27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66bb948c15a1089cc29548df9da9f67b1f554d907a02f6076a0da38cabca9670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66bb948c15a1089cc29548df9da9f67b1f554d907a02f6076a0da38cabca9670\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.357596 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d459b25-158a-4a2f-bff7-b7c643450c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc5172a4b40f579c4ae113095afcdc882c281d31c0ddf2f59f0593c0a8e1837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69b51611cd90f67fb36b564db5890aff54f52a1884d71d297f335d707ec8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5fvqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.368194 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.368233 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.368243 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.368261 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.368272 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:01Z","lastTransitionTime":"2025-11-24T06:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.379954 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"lector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127759 6475 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127828 6475 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127886 6475 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127949 6475 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.128093 6475 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.128496 6475 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 06:55:47.128522 6475 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 06:55:47.128544 6475 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 06:55:47.128555 6475 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 06:55:47.128583 6475 factory.go:656] Stopping watch factory\\\\nI1124 06:55:47.128600 6475 ovnkube.go:599] Stopped ovnkube\\\\nI1124 06:55:47.128621 6475 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kpqhl_openshift-ovn-kubernetes(74622ad8-44b0-44a7-afbd-e19bb9635520)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.391105 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cq8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15c4f8f9-fd68-4636-88cb-d967cb3e202c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d9fa21c921ed12c61aa6a10a4cd4a5c575109a77a6b7099106b76414358d1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbdhx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cq8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.471255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.471297 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.471307 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.471323 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.471336 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:01Z","lastTransitionTime":"2025-11-24T06:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.574410 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.574829 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.574929 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.575054 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.575147 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:01Z","lastTransitionTime":"2025-11-24T06:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.679818 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.679897 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.679915 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.679943 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.679966 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:01Z","lastTransitionTime":"2025-11-24T06:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.788423 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.788528 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.788539 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.788556 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.788566 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:01Z","lastTransitionTime":"2025-11-24T06:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.890579 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.890601 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:01 crc kubenswrapper[4809]: E1124 06:56:01.890827 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:01 crc kubenswrapper[4809]: E1124 06:56:01.891027 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.892314 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.892381 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.892405 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.892436 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.892460 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:01Z","lastTransitionTime":"2025-11-24T06:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.994914 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.994950 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.994959 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.995004 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:01 crc kubenswrapper[4809]: I1124 06:56:01.995017 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:01Z","lastTransitionTime":"2025-11-24T06:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.097407 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.097458 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.097473 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.097491 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.097504 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:02Z","lastTransitionTime":"2025-11-24T06:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.199383 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.199413 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.199420 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.199433 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.199442 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:02Z","lastTransitionTime":"2025-11-24T06:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.301693 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.301720 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.301729 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.301741 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.301749 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:02Z","lastTransitionTime":"2025-11-24T06:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.404642 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.404706 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.404717 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.404739 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.404751 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:02Z","lastTransitionTime":"2025-11-24T06:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.507456 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.507500 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.507511 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.507530 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.507544 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:02Z","lastTransitionTime":"2025-11-24T06:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.609855 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.609894 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.609904 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.609919 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.609929 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:02Z","lastTransitionTime":"2025-11-24T06:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.712260 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.712513 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.712596 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.712681 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.712751 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:02Z","lastTransitionTime":"2025-11-24T06:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.815113 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.815147 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.815159 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.815175 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.815187 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:02Z","lastTransitionTime":"2025-11-24T06:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.890253 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.890281 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:02 crc kubenswrapper[4809]: E1124 06:56:02.890426 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:02 crc kubenswrapper[4809]: E1124 06:56:02.890519 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.917468 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.917561 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.917576 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.917657 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:02 crc kubenswrapper[4809]: I1124 06:56:02.917678 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:02Z","lastTransitionTime":"2025-11-24T06:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.020395 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.020436 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.020447 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.020492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.020505 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:03Z","lastTransitionTime":"2025-11-24T06:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.123066 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.123103 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.123117 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.123136 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.123150 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:03Z","lastTransitionTime":"2025-11-24T06:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.225617 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.225722 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.225744 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.225774 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.225796 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:03Z","lastTransitionTime":"2025-11-24T06:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.329343 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.329416 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.329433 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.329466 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.329492 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:03Z","lastTransitionTime":"2025-11-24T06:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.432074 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.432112 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.432122 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.432137 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.432149 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:03Z","lastTransitionTime":"2025-11-24T06:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.534667 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.534763 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.534884 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.534924 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.534951 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:03Z","lastTransitionTime":"2025-11-24T06:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.637647 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.637706 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.637727 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.637759 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.637782 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:03Z","lastTransitionTime":"2025-11-24T06:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.740284 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.740327 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.740340 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.740357 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.740400 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:03Z","lastTransitionTime":"2025-11-24T06:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.843382 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.843440 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.843456 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.843479 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.843514 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:03Z","lastTransitionTime":"2025-11-24T06:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.890772 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.890862 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:03 crc kubenswrapper[4809]: E1124 06:56:03.890900 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:03 crc kubenswrapper[4809]: E1124 06:56:03.891051 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.945647 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.945691 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.945702 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.945717 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:03 crc kubenswrapper[4809]: I1124 06:56:03.945729 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:03Z","lastTransitionTime":"2025-11-24T06:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.048367 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.048463 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.048482 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.048519 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.048540 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:04Z","lastTransitionTime":"2025-11-24T06:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.151769 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.151817 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.151829 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.151845 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.151858 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:04Z","lastTransitionTime":"2025-11-24T06:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.255257 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.255302 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.255314 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.255333 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.255346 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:04Z","lastTransitionTime":"2025-11-24T06:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.357144 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.357198 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.357216 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.357237 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.357251 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:04Z","lastTransitionTime":"2025-11-24T06:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.459803 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.459868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.459886 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.459911 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.459930 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:04Z","lastTransitionTime":"2025-11-24T06:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.611582 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.611622 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.611633 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.611649 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.611660 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:04Z","lastTransitionTime":"2025-11-24T06:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.714228 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.714261 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.714271 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.714288 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.714299 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:04Z","lastTransitionTime":"2025-11-24T06:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.816356 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.816393 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.816401 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.816415 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.816424 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:04Z","lastTransitionTime":"2025-11-24T06:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.891025 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:04 crc kubenswrapper[4809]: E1124 06:56:04.891209 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.891711 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:04 crc kubenswrapper[4809]: E1124 06:56:04.891880 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.919287 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.919314 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.919323 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.919334 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:04 crc kubenswrapper[4809]: I1124 06:56:04.919343 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:04Z","lastTransitionTime":"2025-11-24T06:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.021299 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.021333 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.021342 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.021394 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.021406 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:05Z","lastTransitionTime":"2025-11-24T06:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.123620 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.123664 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.123672 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.123686 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.123695 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:05Z","lastTransitionTime":"2025-11-24T06:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.225931 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.225990 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.226001 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.226016 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.226028 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:05Z","lastTransitionTime":"2025-11-24T06:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.328498 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.328531 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.328541 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.328555 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.328566 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:05Z","lastTransitionTime":"2025-11-24T06:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.430390 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.430425 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.430436 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.430450 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.430461 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:05Z","lastTransitionTime":"2025-11-24T06:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.532574 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.532608 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.532617 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.532631 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.532642 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:05Z","lastTransitionTime":"2025-11-24T06:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.634789 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.634823 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.634835 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.634850 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.634861 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:05Z","lastTransitionTime":"2025-11-24T06:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.736902 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.737000 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.737014 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.737041 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.737054 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:05Z","lastTransitionTime":"2025-11-24T06:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.839751 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.839817 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.839829 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.839854 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.839868 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:05Z","lastTransitionTime":"2025-11-24T06:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.891140 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.891182 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:05 crc kubenswrapper[4809]: E1124 06:56:05.891261 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:05 crc kubenswrapper[4809]: E1124 06:56:05.891424 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.942347 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.942410 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.942422 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.942439 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:05 crc kubenswrapper[4809]: I1124 06:56:05.942448 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:05Z","lastTransitionTime":"2025-11-24T06:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.044511 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.044552 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.044561 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.044575 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.044585 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:06Z","lastTransitionTime":"2025-11-24T06:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.146756 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.146797 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.146806 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.146824 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.146837 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:06Z","lastTransitionTime":"2025-11-24T06:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.249763 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.249808 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.249818 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.249836 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.249845 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:06Z","lastTransitionTime":"2025-11-24T06:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.353346 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.353395 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.353406 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.353421 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.353431 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:06Z","lastTransitionTime":"2025-11-24T06:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.424190 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs\") pod \"network-metrics-daemon-krmz8\" (UID: \"35b8615e-faa5-47b4-8c22-609d5fda9590\") " pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:06 crc kubenswrapper[4809]: E1124 06:56:06.424820 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:56:06 crc kubenswrapper[4809]: E1124 06:56:06.424886 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs podName:35b8615e-faa5-47b4-8c22-609d5fda9590 nodeName:}" failed. No retries permitted until 2025-11-24 06:56:38.424870985 +0000 UTC m=+98.325462690 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs") pod "network-metrics-daemon-krmz8" (UID: "35b8615e-faa5-47b4-8c22-609d5fda9590") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.455647 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.455699 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.455711 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.455728 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.455739 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:06Z","lastTransitionTime":"2025-11-24T06:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.558341 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.558381 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.558393 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.558409 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.558420 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:06Z","lastTransitionTime":"2025-11-24T06:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.660898 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.660946 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.660958 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.660992 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.661004 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:06Z","lastTransitionTime":"2025-11-24T06:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.763921 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.763958 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.763980 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.763996 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.764006 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:06Z","lastTransitionTime":"2025-11-24T06:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.866535 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.866577 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.866590 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.866608 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.866618 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:06Z","lastTransitionTime":"2025-11-24T06:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.890766 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.890781 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:06 crc kubenswrapper[4809]: E1124 06:56:06.890878 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:06 crc kubenswrapper[4809]: E1124 06:56:06.890979 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.968516 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.968572 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.968588 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.968611 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:06 crc kubenswrapper[4809]: I1124 06:56:06.968628 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:06Z","lastTransitionTime":"2025-11-24T06:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.070650 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.070704 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.070713 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.070727 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.070736 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:07Z","lastTransitionTime":"2025-11-24T06:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.173602 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.173651 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.173661 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.173674 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.173685 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:07Z","lastTransitionTime":"2025-11-24T06:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.276517 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.276614 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.276643 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.276712 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.276737 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:07Z","lastTransitionTime":"2025-11-24T06:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.379746 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.379786 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.379795 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.379810 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.379819 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:07Z","lastTransitionTime":"2025-11-24T06:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.482000 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.482042 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.482050 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.482064 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.482073 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:07Z","lastTransitionTime":"2025-11-24T06:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.583885 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.583945 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.583956 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.583987 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.583997 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:07Z","lastTransitionTime":"2025-11-24T06:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.686555 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.686595 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.686605 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.686619 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.686627 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:07Z","lastTransitionTime":"2025-11-24T06:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.789036 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.789072 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.789082 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.789095 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.789104 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:07Z","lastTransitionTime":"2025-11-24T06:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.890296 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.891012 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:07 crc kubenswrapper[4809]: E1124 06:56:07.891155 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:07 crc kubenswrapper[4809]: E1124 06:56:07.891212 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.891350 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.891403 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.891413 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.891426 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.891435 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:07Z","lastTransitionTime":"2025-11-24T06:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.993711 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.993751 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.993759 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.993774 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.993784 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:07Z","lastTransitionTime":"2025-11-24T06:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.997511 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.997552 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.997564 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.997580 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:07 crc kubenswrapper[4809]: I1124 06:56:07.997590 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:07Z","lastTransitionTime":"2025-11-24T06:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:08 crc kubenswrapper[4809]: E1124 06:56:08.010112 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:08Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.013824 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.013868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.013877 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.013893 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.013903 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:08Z","lastTransitionTime":"2025-11-24T06:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:08 crc kubenswrapper[4809]: E1124 06:56:08.027067 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:08Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.030183 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.030234 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.030245 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.030263 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.030277 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:08Z","lastTransitionTime":"2025-11-24T06:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:08 crc kubenswrapper[4809]: E1124 06:56:08.041991 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:08Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.045350 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.045382 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.045390 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.045404 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.045415 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:08Z","lastTransitionTime":"2025-11-24T06:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:08 crc kubenswrapper[4809]: E1124 06:56:08.055783 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:08Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.059559 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.059609 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.059621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.059635 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.059647 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:08Z","lastTransitionTime":"2025-11-24T06:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:08 crc kubenswrapper[4809]: E1124 06:56:08.070861 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:08Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:08 crc kubenswrapper[4809]: E1124 06:56:08.071000 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.097066 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.097110 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.097122 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.097139 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.097150 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:08Z","lastTransitionTime":"2025-11-24T06:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.199376 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.199422 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.199433 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.199449 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.199461 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:08Z","lastTransitionTime":"2025-11-24T06:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.301871 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.301932 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.301941 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.301957 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.302007 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:08Z","lastTransitionTime":"2025-11-24T06:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.404202 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.404240 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.404249 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.404264 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.404275 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:08Z","lastTransitionTime":"2025-11-24T06:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.506737 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.506773 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.506784 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.506798 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.506806 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:08Z","lastTransitionTime":"2025-11-24T06:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.609452 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.609503 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.609517 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.609555 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.609567 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:08Z","lastTransitionTime":"2025-11-24T06:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.711932 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.712017 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.712029 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.712044 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.712054 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:08Z","lastTransitionTime":"2025-11-24T06:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.814304 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.814366 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.814377 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.814397 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.814409 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:08Z","lastTransitionTime":"2025-11-24T06:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.890863 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.890928 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:08 crc kubenswrapper[4809]: E1124 06:56:08.891001 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:08 crc kubenswrapper[4809]: E1124 06:56:08.891055 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.916414 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.916450 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.916460 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.916474 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:08 crc kubenswrapper[4809]: I1124 06:56:08.916484 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:08Z","lastTransitionTime":"2025-11-24T06:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.018841 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.018879 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.018887 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.018902 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.018911 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:09Z","lastTransitionTime":"2025-11-24T06:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.121730 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.121786 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.121798 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.121816 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.121828 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:09Z","lastTransitionTime":"2025-11-24T06:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.223846 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.223885 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.223893 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.223907 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.223916 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:09Z","lastTransitionTime":"2025-11-24T06:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.326590 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.326804 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.326812 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.326825 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.326834 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:09Z","lastTransitionTime":"2025-11-24T06:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.429202 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.429277 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.429294 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.429317 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.429336 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:09Z","lastTransitionTime":"2025-11-24T06:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.531808 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.531857 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.531868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.531886 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.531900 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:09Z","lastTransitionTime":"2025-11-24T06:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.636143 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.636198 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.636214 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.636231 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.636245 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:09Z","lastTransitionTime":"2025-11-24T06:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.738935 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.739025 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.739046 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.739068 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.739085 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:09Z","lastTransitionTime":"2025-11-24T06:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.841493 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.841549 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.841563 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.841586 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.841602 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:09Z","lastTransitionTime":"2025-11-24T06:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.890620 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.890691 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:09 crc kubenswrapper[4809]: E1124 06:56:09.890768 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:09 crc kubenswrapper[4809]: E1124 06:56:09.890828 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.943756 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.943803 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.943815 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.943831 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:09 crc kubenswrapper[4809]: I1124 06:56:09.943844 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:09Z","lastTransitionTime":"2025-11-24T06:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.034841 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5xvl8_3c932894-56b4-4303-8176-d12b12686b09/kube-multus/0.log" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.034906 4809 generic.go:334] "Generic (PLEG): container finished" podID="3c932894-56b4-4303-8176-d12b12686b09" containerID="0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95" exitCode=1 Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.034941 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5xvl8" event={"ID":"3c932894-56b4-4303-8176-d12b12686b09","Type":"ContainerDied","Data":"0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95"} Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.035473 4809 scope.go:117] "RemoveContainer" containerID="0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.050239 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.050278 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.050288 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.050303 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.050311 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:10Z","lastTransitionTime":"2025-11-24T06:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.055955 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"lector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127759 6475 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127828 6475 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127886 6475 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127949 6475 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.128093 6475 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.128496 6475 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 06:55:47.128522 6475 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 06:55:47.128544 6475 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 06:55:47.128555 6475 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 06:55:47.128583 6475 factory.go:656] Stopping watch factory\\\\nI1124 06:55:47.128600 6475 ovnkube.go:599] Stopped ovnkube\\\\nI1124 06:55:47.128621 6475 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kpqhl_openshift-ovn-kubernetes(74622ad8-44b0-44a7-afbd-e19bb9635520)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.068442 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cq8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15c4f8f9-fd68-4636-88cb-d967cb3e202c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d9fa21c921ed12c61aa6a10a4cd4a5c575109a77a6b7099106b76414358d1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbdhx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cq8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.083008 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.093479 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.106483 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.120578 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:56:09Z\\\",\\\"message\\\":\\\"2025-11-24T06:55:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b6d8a6b1-e4f9-471a-9175-1a14d153febe\\\\n2025-11-24T06:55:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b6d8a6b1-e4f9-471a-9175-1a14d153febe to /host/opt/cni/bin/\\\\n2025-11-24T06:55:23Z [verbose] multus-daemon started\\\\n2025-11-24T06:55:23Z [verbose] Readiness Indicator file check\\\\n2025-11-24T06:56:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.133323 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.145063 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.152328 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.152360 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.152370 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.152385 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.152395 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:10Z","lastTransitionTime":"2025-11-24T06:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.161184 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.174477 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.185109 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krmz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b8615e-faa5-47b4-8c22-609d5fda9590\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krmz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.198261 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.208863 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4e7ae8a-acd5-4685-97b4-9f775d5c08d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80b7097545fca095740245eb7373a8ad9325bfde309725cf0b0db792bbd44df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5376dc11940017714443d1a4cf6c61f323eb35b41d1ce9b76d73137465f18c83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70c7310e5734c08f62e5732d069aed4eb9ccd5cc0e8ae546950871772f6a27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66bb948c15a1089cc29548df9da9f67b1f554d907a02f6076a0da38cabca9670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66bb948c15a1089cc29548df9da9f67b1f554d907a02f6076a0da38cabca9670\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.219715 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.232097 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.243932 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.253776 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.253805 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.253814 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.253826 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.253835 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:10Z","lastTransitionTime":"2025-11-24T06:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.255768 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d459b25-158a-4a2f-bff7-b7c643450c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc5172a4b40f579c4ae113095afcdc882c281d31c0ddf2f59f0593c0a8e1837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69b51611cd90f67fb36b564db5890aff54f52a1884d71d297f335d707ec8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5fvqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.355489 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.355532 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.355542 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.355560 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.355571 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:10Z","lastTransitionTime":"2025-11-24T06:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.457349 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.457380 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.457388 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.457400 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.457408 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:10Z","lastTransitionTime":"2025-11-24T06:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.559837 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.559865 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.559873 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.559885 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.559893 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:10Z","lastTransitionTime":"2025-11-24T06:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.662300 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.662339 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.662347 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.662359 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.662368 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:10Z","lastTransitionTime":"2025-11-24T06:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.764724 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.764760 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.764772 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.764785 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.764795 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:10Z","lastTransitionTime":"2025-11-24T06:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.868431 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.868465 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.868488 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.868512 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.868525 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:10Z","lastTransitionTime":"2025-11-24T06:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.890426 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:10 crc kubenswrapper[4809]: E1124 06:56:10.890548 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.890694 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:10 crc kubenswrapper[4809]: E1124 06:56:10.890737 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.903083 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d459b25-158a-4a2f-bff7-b7c643450c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc5172a4b40f579c4ae113095afcdc882c281d31c0ddf2f59f0593c0a8e1837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69b51611cd90f67fb36b564db5890aff54f52a1884d71d297f335d707ec8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5fvqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.912672 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cq8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15c4f8f9-fd68-4636-88cb-d967cb3e202c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d9fa21c921ed12c61aa6a10a4cd4a5c575109a77a6b7099106b76414358d1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbdhx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cq8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.929027 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"lector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127759 6475 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127828 6475 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127886 6475 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127949 6475 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.128093 6475 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.128496 6475 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 06:55:47.128522 6475 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 06:55:47.128544 6475 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 06:55:47.128555 6475 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 06:55:47.128583 6475 factory.go:656] Stopping watch factory\\\\nI1124 06:55:47.128600 6475 ovnkube.go:599] Stopped ovnkube\\\\nI1124 06:55:47.128621 6475 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kpqhl_openshift-ovn-kubernetes(74622ad8-44b0-44a7-afbd-e19bb9635520)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.939693 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.950372 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.960671 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.971461 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.971523 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.971540 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.971563 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.971579 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:10Z","lastTransitionTime":"2025-11-24T06:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.971895 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.983618 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:56:09Z\\\",\\\"message\\\":\\\"2025-11-24T06:55:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b6d8a6b1-e4f9-471a-9175-1a14d153febe\\\\n2025-11-24T06:55:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b6d8a6b1-e4f9-471a-9175-1a14d153febe to /host/opt/cni/bin/\\\\n2025-11-24T06:55:23Z [verbose] multus-daemon started\\\\n2025-11-24T06:55:23Z [verbose] Readiness Indicator file check\\\\n2025-11-24T06:56:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:10 crc kubenswrapper[4809]: I1124 06:56:10.998378 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.010436 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.023121 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.032500 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krmz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b8615e-faa5-47b4-8c22-609d5fda9590\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krmz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.039701 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5xvl8_3c932894-56b4-4303-8176-d12b12686b09/kube-multus/0.log" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.039750 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5xvl8" event={"ID":"3c932894-56b4-4303-8176-d12b12686b09","Type":"ContainerStarted","Data":"d68cb2e2c7982f3c55f7d9263588e648ba155a930efc0266051aaa628ccb9db7"} Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.044396 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.057068 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.066684 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4e7ae8a-acd5-4685-97b4-9f775d5c08d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80b7097545fca095740245eb7373a8ad9325bfde309725cf0b0db792bbd44df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5376dc11940017714443d1a4cf6c61f323eb35b41d1ce9b76d73137465f18c83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70c7310e5734c08f62e5732d069aed4eb9ccd5cc0e8ae546950871772f6a27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66bb948c15a1089cc29548df9da9f67b1f554d907a02f6076a0da38cabca9670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66bb948c15a1089cc29548df9da9f67b1f554d907a02f6076a0da38cabca9670\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.075644 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.075699 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.075763 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.075867 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.075949 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:11Z","lastTransitionTime":"2025-11-24T06:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.077909 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.089556 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.105739 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"lector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127759 6475 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127828 6475 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127886 6475 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127949 6475 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.128093 6475 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.128496 6475 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 06:55:47.128522 6475 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 06:55:47.128544 6475 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 06:55:47.128555 6475 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 06:55:47.128583 6475 factory.go:656] Stopping watch factory\\\\nI1124 06:55:47.128600 6475 ovnkube.go:599] Stopped ovnkube\\\\nI1124 06:55:47.128621 6475 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kpqhl_openshift-ovn-kubernetes(74622ad8-44b0-44a7-afbd-e19bb9635520)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.115080 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cq8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15c4f8f9-fd68-4636-88cb-d967cb3e202c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d9fa21c921ed12c61aa6a10a4cd4a5c575109a77a6b7099106b76414358d1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbdhx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cq8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.125839 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68cb2e2c7982f3c55f7d9263588e648ba155a930efc0266051aaa628ccb9db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:56:09Z\\\",\\\"message\\\":\\\"2025-11-24T06:55:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b6d8a6b1-e4f9-471a-9175-1a14d153febe\\\\n2025-11-24T06:55:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b6d8a6b1-e4f9-471a-9175-1a14d153febe to /host/opt/cni/bin/\\\\n2025-11-24T06:55:23Z [verbose] multus-daemon started\\\\n2025-11-24T06:55:23Z [verbose] Readiness Indicator file check\\\\n2025-11-24T06:56:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.136158 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.146696 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.159613 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.171921 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.179112 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.179139 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.179148 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.179161 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.179170 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:11Z","lastTransitionTime":"2025-11-24T06:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.184016 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.193019 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.201836 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.211441 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krmz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b8615e-faa5-47b4-8c22-609d5fda9590\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krmz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.222861 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4e7ae8a-acd5-4685-97b4-9f775d5c08d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80b7097545fca095740245eb7373a8ad9325bfde309725cf0b0db792bbd44df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5376dc11940017714443d1a4cf6c61f323eb35b41d1ce9b76d73137465f18c83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70c7310e5734c08f62e5732d069aed4eb9ccd5cc0e8ae546950871772f6a27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66bb948c15a1089cc29548df9da9f67b1f554d907a02f6076a0da38cabca9670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66bb948c15a1089cc29548df9da9f67b1f554d907a02f6076a0da38cabca9670\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.236133 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.248410 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.260446 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.275229 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.281621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.281694 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.281706 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.281723 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.281734 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:11Z","lastTransitionTime":"2025-11-24T06:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.285105 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d459b25-158a-4a2f-bff7-b7c643450c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc5172a4b40f579c4ae113095afcdc882c281d31c0ddf2f59f0593c0a8e1837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69b51611cd90f67fb36b564db5890aff54f52a1884d71d297f335d707ec8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5fvqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.383887 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.383923 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.383934 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.384014 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.384043 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:11Z","lastTransitionTime":"2025-11-24T06:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.486540 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.486577 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.486587 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.486600 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.486610 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:11Z","lastTransitionTime":"2025-11-24T06:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.588859 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.588898 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.588908 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.588925 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.588938 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:11Z","lastTransitionTime":"2025-11-24T06:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.692255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.692287 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.692297 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.692312 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.692322 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:11Z","lastTransitionTime":"2025-11-24T06:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.794737 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.794777 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.794788 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.794802 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.794811 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:11Z","lastTransitionTime":"2025-11-24T06:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.891122 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.891135 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:11 crc kubenswrapper[4809]: E1124 06:56:11.891265 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:11 crc kubenswrapper[4809]: E1124 06:56:11.891382 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.896525 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.896558 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.896569 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.896586 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.896597 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:11Z","lastTransitionTime":"2025-11-24T06:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.998397 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.998444 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.998452 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.998464 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:11 crc kubenswrapper[4809]: I1124 06:56:11.998473 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:11Z","lastTransitionTime":"2025-11-24T06:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.100877 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.100913 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.100922 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.100936 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.100946 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:12Z","lastTransitionTime":"2025-11-24T06:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.202538 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.202566 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.202576 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.202590 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.202599 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:12Z","lastTransitionTime":"2025-11-24T06:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.304924 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.305021 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.305038 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.305062 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.305080 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:12Z","lastTransitionTime":"2025-11-24T06:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.406731 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.406766 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.406777 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.406795 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.406806 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:12Z","lastTransitionTime":"2025-11-24T06:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.509356 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.509420 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.509443 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.509472 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.509493 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:12Z","lastTransitionTime":"2025-11-24T06:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.611455 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.611524 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.611547 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.611569 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.611586 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:12Z","lastTransitionTime":"2025-11-24T06:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.713898 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.713932 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.713939 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.713952 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.713960 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:12Z","lastTransitionTime":"2025-11-24T06:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.816162 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.816219 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.816230 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.816246 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.816257 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:12Z","lastTransitionTime":"2025-11-24T06:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.890171 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.890230 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:12 crc kubenswrapper[4809]: E1124 06:56:12.890324 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:12 crc kubenswrapper[4809]: E1124 06:56:12.890383 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.891147 4809 scope.go:117] "RemoveContainer" containerID="1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.918727 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.918762 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.918773 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.918786 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:12 crc kubenswrapper[4809]: I1124 06:56:12.918796 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:12Z","lastTransitionTime":"2025-11-24T06:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.021660 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.021698 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.021707 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.021721 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.021730 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:13Z","lastTransitionTime":"2025-11-24T06:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.046550 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kpqhl_74622ad8-44b0-44a7-afbd-e19bb9635520/ovnkube-controller/2.log" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.049140 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerStarted","Data":"7f3517cb223a94e59385d72b6844b837a34cb285128c89db3ef6f259a4d505f3"} Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.050073 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.064387 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d459b25-158a-4a2f-bff7-b7c643450c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc5172a4b40f579c4ae113095afcdc882c281d31c0ddf2f59f0593c0a8e1837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69b51611cd90f67fb36b564db5890aff54f52a1884d71d297f335d707ec8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5fvqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:13Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.082069 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3517cb223a94e59385d72b6844b837a34cb285128c89db3ef6f259a4d505f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"lector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127759 6475 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127828 6475 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127886 6475 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127949 6475 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.128093 6475 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.128496 6475 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 06:55:47.128522 6475 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 06:55:47.128544 6475 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 06:55:47.128555 6475 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 06:55:47.128583 6475 factory.go:656] Stopping watch factory\\\\nI1124 06:55:47.128600 6475 ovnkube.go:599] Stopped ovnkube\\\\nI1124 06:55:47.128621 6475 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:13Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.090879 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cq8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15c4f8f9-fd68-4636-88cb-d967cb3e202c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d9fa21c921ed12c61aa6a10a4cd4a5c575109a77a6b7099106b76414358d1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbdhx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cq8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:13Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.101741 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68cb2e2c7982f3c55f7d9263588e648ba155a930efc0266051aaa628ccb9db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:56:09Z\\\",\\\"message\\\":\\\"2025-11-24T06:55:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b6d8a6b1-e4f9-471a-9175-1a14d153febe\\\\n2025-11-24T06:55:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b6d8a6b1-e4f9-471a-9175-1a14d153febe to /host/opt/cni/bin/\\\\n2025-11-24T06:55:23Z [verbose] multus-daemon started\\\\n2025-11-24T06:55:23Z [verbose] Readiness Indicator file check\\\\n2025-11-24T06:56:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:13Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.117853 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:13Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.123441 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.123474 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.123485 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.123499 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.123509 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:13Z","lastTransitionTime":"2025-11-24T06:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.129234 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:13Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.141374 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:13Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.152756 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:13Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.164490 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:13Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.175079 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:13Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.187598 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:13Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.198655 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krmz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b8615e-faa5-47b4-8c22-609d5fda9590\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krmz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:13Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.216044 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4e7ae8a-acd5-4685-97b4-9f775d5c08d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80b7097545fca095740245eb7373a8ad9325bfde309725cf0b0db792bbd44df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5376dc11940017714443d1a4cf6c61f323eb35b41d1ce9b76d73137465f18c83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70c7310e5734c08f62e5732d069aed4eb9ccd5cc0e8ae546950871772f6a27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66bb948c15a1089cc29548df9da9f67b1f554d907a02f6076a0da38cabca9670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66bb948c15a1089cc29548df9da9f67b1f554d907a02f6076a0da38cabca9670\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:13Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.225484 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.225508 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.225517 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.225530 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.225539 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:13Z","lastTransitionTime":"2025-11-24T06:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.231396 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:13Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.244520 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:13Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.267735 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:13Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.294892 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:13Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.327221 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.327264 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.327274 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.327290 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.327300 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:13Z","lastTransitionTime":"2025-11-24T06:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.429300 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.429335 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.429346 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.429360 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.429371 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:13Z","lastTransitionTime":"2025-11-24T06:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.531811 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.531861 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.531877 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.531894 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.531907 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:13Z","lastTransitionTime":"2025-11-24T06:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.634710 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.634747 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.634756 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.634769 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.634778 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:13Z","lastTransitionTime":"2025-11-24T06:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.736903 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.737029 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.737046 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.737061 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.737072 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:13Z","lastTransitionTime":"2025-11-24T06:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.840354 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.840394 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.840402 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.840415 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.840428 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:13Z","lastTransitionTime":"2025-11-24T06:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.891143 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.891167 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:13 crc kubenswrapper[4809]: E1124 06:56:13.891289 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:13 crc kubenswrapper[4809]: E1124 06:56:13.891396 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.943026 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.943074 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.943087 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.943101 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:13 crc kubenswrapper[4809]: I1124 06:56:13.943113 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:13Z","lastTransitionTime":"2025-11-24T06:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.045810 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.045852 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.045862 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.045877 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.045887 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:14Z","lastTransitionTime":"2025-11-24T06:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.053457 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kpqhl_74622ad8-44b0-44a7-afbd-e19bb9635520/ovnkube-controller/3.log" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.053902 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kpqhl_74622ad8-44b0-44a7-afbd-e19bb9635520/ovnkube-controller/2.log" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.056559 4809 generic.go:334] "Generic (PLEG): container finished" podID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerID="7f3517cb223a94e59385d72b6844b837a34cb285128c89db3ef6f259a4d505f3" exitCode=1 Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.056587 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerDied","Data":"7f3517cb223a94e59385d72b6844b837a34cb285128c89db3ef6f259a4d505f3"} Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.056616 4809 scope.go:117] "RemoveContainer" containerID="1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.057256 4809 scope.go:117] "RemoveContainer" containerID="7f3517cb223a94e59385d72b6844b837a34cb285128c89db3ef6f259a4d505f3" Nov 24 06:56:14 crc kubenswrapper[4809]: E1124 06:56:14.057422 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-kpqhl_openshift-ovn-kubernetes(74622ad8-44b0-44a7-afbd-e19bb9635520)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.075910 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3517cb223a94e59385d72b6844b837a34cb285128c89db3ef6f259a4d505f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a82317bec8c525f2027d59c0d63b2b74c54ffa3324634397fcb7a13375235fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:55:47Z\\\",\\\"message\\\":\\\"lector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127759 6475 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127828 6475 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127886 6475 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.127949 6475 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.128093 6475 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:55:47.128496 6475 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 06:55:47.128522 6475 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 06:55:47.128544 6475 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 06:55:47.128555 6475 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 06:55:47.128583 6475 factory.go:656] Stopping watch factory\\\\nI1124 06:55:47.128600 6475 ovnkube.go:599] Stopped ovnkube\\\\nI1124 06:55:47.128621 6475 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f3517cb223a94e59385d72b6844b837a34cb285128c89db3ef6f259a4d505f3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:56:13Z\\\",\\\"message\\\":\\\" object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-dr8hv\\\\nI1124 06:56:13.635999 6838 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-dr8hv in node crc\\\\nI1124 06:56:13.635885 6838 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1124 06:56:13.636009 6838 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-dr8hv after 0 failed attempt(s)\\\\nI1124 06:56:13.636017 6838 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-dr8hv\\\\nI1124 06:56:13.635801 6838 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 06:56:13.636017 6838 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI1124 06:56:13.636110 6838 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nF1124 06:56:13.636122 6838 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:14Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.085830 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cq8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15c4f8f9-fd68-4636-88cb-d967cb3e202c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d9fa21c921ed12c61aa6a10a4cd4a5c575109a77a6b7099106b76414358d1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbdhx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cq8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:14Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.100652 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:14Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.111373 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:14Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.122343 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:14Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.132143 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:14Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.157885 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.157941 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.157956 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.157998 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.158021 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:14Z","lastTransitionTime":"2025-11-24T06:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.169601 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:14Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.187470 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:14Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.202529 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:14Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.216581 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68cb2e2c7982f3c55f7d9263588e648ba155a930efc0266051aaa628ccb9db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:56:09Z\\\",\\\"message\\\":\\\"2025-11-24T06:55:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b6d8a6b1-e4f9-471a-9175-1a14d153febe\\\\n2025-11-24T06:55:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b6d8a6b1-e4f9-471a-9175-1a14d153febe to /host/opt/cni/bin/\\\\n2025-11-24T06:55:23Z [verbose] multus-daemon started\\\\n2025-11-24T06:55:23Z [verbose] Readiness Indicator file check\\\\n2025-11-24T06:56:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:14Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.225200 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krmz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b8615e-faa5-47b4-8c22-609d5fda9590\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krmz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:14Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.234675 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4e7ae8a-acd5-4685-97b4-9f775d5c08d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80b7097545fca095740245eb7373a8ad9325bfde309725cf0b0db792bbd44df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5376dc11940017714443d1a4cf6c61f323eb35b41d1ce9b76d73137465f18c83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70c7310e5734c08f62e5732d069aed4eb9ccd5cc0e8ae546950871772f6a27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66bb948c15a1089cc29548df9da9f67b1f554d907a02f6076a0da38cabca9670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66bb948c15a1089cc29548df9da9f67b1f554d907a02f6076a0da38cabca9670\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:14Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.244881 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:14Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.255925 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:14Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.260072 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.260106 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.260113 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.260127 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.260136 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:14Z","lastTransitionTime":"2025-11-24T06:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.266122 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:14Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.278569 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:14Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.289643 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d459b25-158a-4a2f-bff7-b7c643450c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc5172a4b40f579c4ae113095afcdc882c281d31c0ddf2f59f0593c0a8e1837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69b51611cd90f67fb36b564db5890aff54f52a1884d71d297f335d707ec8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5fvqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:14Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.362835 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.362895 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.362914 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.362939 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.362956 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:14Z","lastTransitionTime":"2025-11-24T06:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.465843 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.465883 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.465895 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.465910 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.465923 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:14Z","lastTransitionTime":"2025-11-24T06:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.568469 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.568499 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.568507 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.568519 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.568527 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:14Z","lastTransitionTime":"2025-11-24T06:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.670605 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.670683 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.670701 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.670726 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.670742 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:14Z","lastTransitionTime":"2025-11-24T06:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.773807 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.773854 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.773863 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.773878 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.773888 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:14Z","lastTransitionTime":"2025-11-24T06:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.876794 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.876858 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.876876 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.876902 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.876922 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:14Z","lastTransitionTime":"2025-11-24T06:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.890490 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:14 crc kubenswrapper[4809]: E1124 06:56:14.890608 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.890491 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:14 crc kubenswrapper[4809]: E1124 06:56:14.890678 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.906060 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.979741 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.980117 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.980275 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.980488 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:14 crc kubenswrapper[4809]: I1124 06:56:14.980585 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:14Z","lastTransitionTime":"2025-11-24T06:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.062000 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kpqhl_74622ad8-44b0-44a7-afbd-e19bb9635520/ovnkube-controller/3.log" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.066003 4809 scope.go:117] "RemoveContainer" containerID="7f3517cb223a94e59385d72b6844b837a34cb285128c89db3ef6f259a4d505f3" Nov 24 06:56:15 crc kubenswrapper[4809]: E1124 06:56:15.066196 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-kpqhl_openshift-ovn-kubernetes(74622ad8-44b0-44a7-afbd-e19bb9635520)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.082088 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4e7ae8a-acd5-4685-97b4-9f775d5c08d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80b7097545fca095740245eb7373a8ad9325bfde309725cf0b0db792bbd44df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5376dc11940017714443d1a4cf6c61f323eb35b41d1ce9b76d73137465f18c83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70c7310e5734c08f62e5732d069aed4eb9ccd5cc0e8ae546950871772f6a27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66bb948c15a1089cc29548df9da9f67b1f554d907a02f6076a0da38cabca9670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66bb948c15a1089cc29548df9da9f67b1f554d907a02f6076a0da38cabca9670\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.085688 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.085745 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.085770 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.085810 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.085835 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:15Z","lastTransitionTime":"2025-11-24T06:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.102922 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.120863 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.135884 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.150509 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.162215 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d459b25-158a-4a2f-bff7-b7c643450c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc5172a4b40f579c4ae113095afcdc882c281d31c0ddf2f59f0593c0a8e1837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69b51611cd90f67fb36b564db5890aff54f52a1884d71d297f335d707ec8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5fvqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.173002 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2345f70b-8c8a-48c3-97cc-c08fbfe61d77\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e21b2b68cf22971d6cedd4d455d80409d8d00e569d5b2ce0319b15e79c91d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb1a74c3f4524f23702cf96f82ca151699897f7a1fe9eafdb035962e93ed897d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb1a74c3f4524f23702cf96f82ca151699897f7a1fe9eafdb035962e93ed897d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.188102 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.188203 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.188217 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.188251 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.188261 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:15Z","lastTransitionTime":"2025-11-24T06:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.194411 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3517cb223a94e59385d72b6844b837a34cb285128c89db3ef6f259a4d505f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f3517cb223a94e59385d72b6844b837a34cb285128c89db3ef6f259a4d505f3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:56:13Z\\\",\\\"message\\\":\\\" object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-dr8hv\\\\nI1124 06:56:13.635999 6838 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-dr8hv in node crc\\\\nI1124 06:56:13.635885 6838 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1124 06:56:13.636009 6838 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-dr8hv after 0 failed attempt(s)\\\\nI1124 06:56:13.636017 6838 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-dr8hv\\\\nI1124 06:56:13.635801 6838 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 06:56:13.636017 6838 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI1124 06:56:13.636110 6838 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nF1124 06:56:13.636122 6838 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:56:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-kpqhl_openshift-ovn-kubernetes(74622ad8-44b0-44a7-afbd-e19bb9635520)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.207352 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cq8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15c4f8f9-fd68-4636-88cb-d967cb3e202c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d9fa21c921ed12c61aa6a10a4cd4a5c575109a77a6b7099106b76414358d1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbdhx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cq8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.221882 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68cb2e2c7982f3c55f7d9263588e648ba155a930efc0266051aaa628ccb9db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:56:09Z\\\",\\\"message\\\":\\\"2025-11-24T06:55:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b6d8a6b1-e4f9-471a-9175-1a14d153febe\\\\n2025-11-24T06:55:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b6d8a6b1-e4f9-471a-9175-1a14d153febe to /host/opt/cni/bin/\\\\n2025-11-24T06:55:23Z [verbose] multus-daemon started\\\\n2025-11-24T06:55:23Z [verbose] Readiness Indicator file check\\\\n2025-11-24T06:56:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.234801 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.245648 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.257167 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.267936 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.280499 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.290583 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.290751 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.290822 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.290882 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.290945 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:15Z","lastTransitionTime":"2025-11-24T06:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.291021 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.300840 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.310822 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krmz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b8615e-faa5-47b4-8c22-609d5fda9590\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krmz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.393235 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.393275 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.393287 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.393304 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.393317 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:15Z","lastTransitionTime":"2025-11-24T06:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.495192 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.495222 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.495230 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.495243 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.495252 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:15Z","lastTransitionTime":"2025-11-24T06:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.598440 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.598469 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.598477 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.598488 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.598497 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:15Z","lastTransitionTime":"2025-11-24T06:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.700547 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.700603 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.700620 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.700642 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.700660 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:15Z","lastTransitionTime":"2025-11-24T06:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.802915 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.802955 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.802977 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.802992 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.803000 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:15Z","lastTransitionTime":"2025-11-24T06:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.890703 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.890750 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:15 crc kubenswrapper[4809]: E1124 06:56:15.890847 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:15 crc kubenswrapper[4809]: E1124 06:56:15.891024 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.906171 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.906218 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.906234 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.906259 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:15 crc kubenswrapper[4809]: I1124 06:56:15.906277 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:15Z","lastTransitionTime":"2025-11-24T06:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.008635 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.008671 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.008679 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.008691 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.008700 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:16Z","lastTransitionTime":"2025-11-24T06:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.111426 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.111591 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.111614 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.111639 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.111663 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:16Z","lastTransitionTime":"2025-11-24T06:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.214109 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.214197 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.214210 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.214226 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.214630 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:16Z","lastTransitionTime":"2025-11-24T06:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.316544 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.316598 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.316615 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.316641 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.316660 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:16Z","lastTransitionTime":"2025-11-24T06:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.418887 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.418936 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.418958 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.419007 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.419024 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:16Z","lastTransitionTime":"2025-11-24T06:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.522142 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.522174 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.522182 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.522195 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.522204 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:16Z","lastTransitionTime":"2025-11-24T06:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.626014 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.626047 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.626079 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.626096 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.626106 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:16Z","lastTransitionTime":"2025-11-24T06:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.728089 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.728148 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.728161 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.728176 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.728186 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:16Z","lastTransitionTime":"2025-11-24T06:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.831014 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.831102 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.831121 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.831146 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.831162 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:16Z","lastTransitionTime":"2025-11-24T06:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.890780 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:16 crc kubenswrapper[4809]: E1124 06:56:16.891074 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.890780 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:16 crc kubenswrapper[4809]: E1124 06:56:16.891584 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.934060 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.934114 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.934130 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.934154 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:16 crc kubenswrapper[4809]: I1124 06:56:16.934172 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:16Z","lastTransitionTime":"2025-11-24T06:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.036368 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.036904 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.036953 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.037189 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.037207 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:17Z","lastTransitionTime":"2025-11-24T06:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.140769 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.140844 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.140856 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.140873 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.140885 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:17Z","lastTransitionTime":"2025-11-24T06:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.244718 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.245296 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.245601 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.245816 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.246064 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:17Z","lastTransitionTime":"2025-11-24T06:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.349888 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.349958 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.350075 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.350107 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.350129 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:17Z","lastTransitionTime":"2025-11-24T06:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.452722 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.453252 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.453455 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.453821 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.454068 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:17Z","lastTransitionTime":"2025-11-24T06:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.557283 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.557320 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.557347 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.557360 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.557368 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:17Z","lastTransitionTime":"2025-11-24T06:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.659406 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.659476 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.659493 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.659519 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.659537 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:17Z","lastTransitionTime":"2025-11-24T06:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.762172 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.762212 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.762224 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.762241 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.762253 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:17Z","lastTransitionTime":"2025-11-24T06:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.865301 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.865362 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.865379 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.865404 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.865420 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:17Z","lastTransitionTime":"2025-11-24T06:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.890615 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.890681 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:17 crc kubenswrapper[4809]: E1124 06:56:17.891078 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:17 crc kubenswrapper[4809]: E1124 06:56:17.891225 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.967566 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.967784 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.967844 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.967947 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:17 crc kubenswrapper[4809]: I1124 06:56:17.968039 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:17Z","lastTransitionTime":"2025-11-24T06:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.070531 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.070590 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.070607 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.070624 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.070636 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:18Z","lastTransitionTime":"2025-11-24T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.118868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.118909 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.118920 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.118936 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.118947 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:18Z","lastTransitionTime":"2025-11-24T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:18 crc kubenswrapper[4809]: E1124 06:56:18.129866 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:18Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.132707 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.132730 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.132738 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.132769 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.132777 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:18Z","lastTransitionTime":"2025-11-24T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:18 crc kubenswrapper[4809]: E1124 06:56:18.142872 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:18Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.145905 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.145934 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.145943 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.145958 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.145989 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:18Z","lastTransitionTime":"2025-11-24T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:18 crc kubenswrapper[4809]: E1124 06:56:18.161274 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:18Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.164774 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.165003 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.165129 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.165244 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.165343 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:18Z","lastTransitionTime":"2025-11-24T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:18 crc kubenswrapper[4809]: E1124 06:56:18.180900 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:18Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.185468 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.185514 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.185530 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.185548 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.185559 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:18Z","lastTransitionTime":"2025-11-24T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:18 crc kubenswrapper[4809]: E1124 06:56:18.198629 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f7265f8-062f-41c9-817e-b7f54d138dae\\\",\\\"systemUUID\\\":\\\"c636b03a-5cf2-4b68-8edd-3198ca3d19e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:18Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:18 crc kubenswrapper[4809]: E1124 06:56:18.198745 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.200283 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.200331 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.200343 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.200357 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.200366 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:18Z","lastTransitionTime":"2025-11-24T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.302296 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.302329 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.302337 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.302350 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.302359 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:18Z","lastTransitionTime":"2025-11-24T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.404826 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.404873 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.404885 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.404902 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.404913 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:18Z","lastTransitionTime":"2025-11-24T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.506880 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.506915 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.506927 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.506943 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.506955 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:18Z","lastTransitionTime":"2025-11-24T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.609071 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.609113 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.609122 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.609137 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.609147 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:18Z","lastTransitionTime":"2025-11-24T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.712074 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.712314 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.712388 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.712458 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.712527 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:18Z","lastTransitionTime":"2025-11-24T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.815500 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.815832 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.816001 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.816167 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.816303 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:18Z","lastTransitionTime":"2025-11-24T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.890809 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:18 crc kubenswrapper[4809]: E1124 06:56:18.891076 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.890824 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:18 crc kubenswrapper[4809]: E1124 06:56:18.891487 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.918847 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.918922 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.918945 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.919005 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:18 crc kubenswrapper[4809]: I1124 06:56:18.919030 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:18Z","lastTransitionTime":"2025-11-24T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.021550 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.022014 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.022235 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.022473 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.022667 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:19Z","lastTransitionTime":"2025-11-24T06:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.126267 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.126333 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.126355 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.126384 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.126407 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:19Z","lastTransitionTime":"2025-11-24T06:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.228714 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.228751 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.228759 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.228773 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.228781 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:19Z","lastTransitionTime":"2025-11-24T06:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.331849 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.331890 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.331901 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.331918 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.331931 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:19Z","lastTransitionTime":"2025-11-24T06:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.434699 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.434750 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.434763 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.434784 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.434798 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:19Z","lastTransitionTime":"2025-11-24T06:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.537841 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.538197 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.538305 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.538402 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.538488 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:19Z","lastTransitionTime":"2025-11-24T06:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.641954 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.642026 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.642037 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.642056 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.642068 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:19Z","lastTransitionTime":"2025-11-24T06:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.744907 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.744960 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.744997 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.745014 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.745033 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:19Z","lastTransitionTime":"2025-11-24T06:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.848406 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.848436 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.848444 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.848456 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.848464 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:19Z","lastTransitionTime":"2025-11-24T06:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.891211 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.891344 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:19 crc kubenswrapper[4809]: E1124 06:56:19.892331 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:19 crc kubenswrapper[4809]: E1124 06:56:19.892441 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.951422 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.951505 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.951532 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.951586 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:19 crc kubenswrapper[4809]: I1124 06:56:19.951612 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:19Z","lastTransitionTime":"2025-11-24T06:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.055440 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.055725 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.055821 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.055928 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.056073 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:20Z","lastTransitionTime":"2025-11-24T06:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.158388 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.158779 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.158929 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.159157 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.159353 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:20Z","lastTransitionTime":"2025-11-24T06:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.263561 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.263632 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.263651 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.263680 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.263699 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:20Z","lastTransitionTime":"2025-11-24T06:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.365995 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.366050 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.366063 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.366080 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.366091 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:20Z","lastTransitionTime":"2025-11-24T06:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.473759 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.474468 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.474493 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.474522 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.474543 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:20Z","lastTransitionTime":"2025-11-24T06:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.578213 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.578272 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.578293 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.578321 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.578343 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:20Z","lastTransitionTime":"2025-11-24T06:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.682204 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.682523 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.682674 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.682814 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.682944 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:20Z","lastTransitionTime":"2025-11-24T06:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.786177 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.786553 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.786719 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.786898 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.787110 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:20Z","lastTransitionTime":"2025-11-24T06:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.890176 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:20 crc kubenswrapper[4809]: E1124 06:56:20.890357 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.890432 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.890550 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.890597 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.890609 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:20 crc kubenswrapper[4809]: E1124 06:56:20.890601 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.890622 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.890662 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:20Z","lastTransitionTime":"2025-11-24T06:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.915526 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74622ad8-44b0-44a7-afbd-e19bb9635520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3517cb223a94e59385d72b6844b837a34cb285128c89db3ef6f259a4d505f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f3517cb223a94e59385d72b6844b837a34cb285128c89db3ef6f259a4d505f3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:56:13Z\\\",\\\"message\\\":\\\" object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-dr8hv\\\\nI1124 06:56:13.635999 6838 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-dr8hv in node crc\\\\nI1124 06:56:13.635885 6838 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1124 06:56:13.636009 6838 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-dr8hv after 0 failed attempt(s)\\\\nI1124 06:56:13.636017 6838 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-dr8hv\\\\nI1124 06:56:13.635801 6838 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 06:56:13.636017 6838 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI1124 06:56:13.636110 6838 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nF1124 06:56:13.636122 6838 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:56:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-kpqhl_openshift-ovn-kubernetes(74622ad8-44b0-44a7-afbd-e19bb9635520)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7hjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kpqhl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:20Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.931026 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cq8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15c4f8f9-fd68-4636-88cb-d967cb3e202c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d9fa21c921ed12c61aa6a10a4cd4a5c575109a77a6b7099106b76414358d1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbdhx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cq8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:20Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.945484 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2345f70b-8c8a-48c3-97cc-c08fbfe61d77\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e21b2b68cf22971d6cedd4d455d80409d8d00e569d5b2ce0319b15e79c91d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb1a74c3f4524f23702cf96f82ca151699897f7a1fe9eafdb035962e93ed897d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb1a74c3f4524f23702cf96f82ca151699897f7a1fe9eafdb035962e93ed897d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:20Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.962826 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e260625758943a1d954ad8cee5dfa12520853043623e54f52b9009306b617d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:20Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.976792 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:20Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.991215 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63be3c7f71af7c556875f1e36c22f1201a5ed1bf58e72abab2758eb470446c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03f4b077b84289f4e1e53cc9efa2b912ddc4d9139e6fdfe796730c4936168446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:20Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.992460 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.992511 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.992525 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.992541 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:20 crc kubenswrapper[4809]: I1124 06:56:20.992555 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:20Z","lastTransitionTime":"2025-11-24T06:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.001920 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8p7tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8cb5c-3973-4f32-984b-5e1c7b4e9f09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50c458d97472a03cccdfdac9175c1364bca9665134f8ac63d1b7717d59e450c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62n55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8p7tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:21Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.012783 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8050d5-1aea-453e-a06e-87702b0e856b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a183ddd998c723d98f42eb5111c7d7de63d2ca5c88e04c7ef779c959a8d46e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sl4qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dr8hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:21Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.031644 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5xvl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c932894-56b4-4303-8176-d12b12686b09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68cb2e2c7982f3c55f7d9263588e648ba155a930efc0266051aaa628ccb9db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:56:09Z\\\",\\\"message\\\":\\\"2025-11-24T06:55:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b6d8a6b1-e4f9-471a-9175-1a14d153febe\\\\n2025-11-24T06:55:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b6d8a6b1-e4f9-471a-9175-1a14d153febe to /host/opt/cni/bin/\\\\n2025-11-24T06:55:23Z [verbose] multus-daemon started\\\\n2025-11-24T06:55:23Z [verbose] Readiness Indicator file check\\\\n2025-11-24T06:56:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gs7tj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5xvl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:21Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.046944 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e68f195e-8670-4b46-abfe-f838323c656b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b390d4238bbbc48ffba690461d55d9f8490eba094c9ac13068c5218c3455cbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1896314c2c439f2f92944f65e9ac1f5a760a22d7860201689fa062cc462b11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612ae5f0191cc5c131f34f854362cb1c82de9b9c7d043709bc1f40fb784a18b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f0c2089a0476932bc9caff63a2364a23aac879425b8fd762e02f8182a2b92d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://854ffdc02077c93dcd334b370528ac66224d703276b6234b3079ad6f4ae06aa2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:55:14.730162 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:55:14.730923 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3030451173/tls.crt::/tmp/serving-cert-3030451173/tls.key\\\\\\\"\\\\nI1124 06:55:20.139882 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:55:20.143286 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:55:20.143316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:55:20.143353 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:55:20.143363 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:55:20.150154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:55:20.150188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150199 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:55:20.150207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:55:20.150212 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:55:20.150217 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:55:20.150221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:55:20.150427 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:55:20.151698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5314f3677007cd03d077af2555c75c5cc111f988616b5947a23dff5247e4c50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386ed8032bfcf9e2339ca20a3ae1a72e852ab1ced2686bd2d6f7678dd6ed00a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:21Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.059511 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:21Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.068741 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krmz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b8615e-faa5-47b4-8c22-609d5fda9590\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-76zdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krmz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:21Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.079139 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:21Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.088903 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eaec82d2c3b91da59be749ee5a5086297e17b81fd310ee3d0bdf44558e1b675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:21Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.094646 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.094698 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.094707 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.094722 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.094732 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:21Z","lastTransitionTime":"2025-11-24T06:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.101498 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86405460-8db2-444c-8273-f2e7040733d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f31660d37ff0aa68455bb55d6f6358f499600b0b225c2fd9d679b6e2db794541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209e232474f576572d17c13da69ea2ba88478283ac47433e0f3396f94c5996d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a113edfec63146f10dab394c943aa3d3177da9e27fd5a310cb2d7edd653d0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5c38a5bfd71084223df05f5dc3bee5b374b25c4cc0d37e9a9b7b7dd45a5d7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://795f6554f8d095aee7d278ca71d744a37f686f0026f660465d6c828c317d4c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a4137d2f41eb7b88981de18ec3c3942d04beda3d121709d1f02c3b47957882b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4dee29b2ebe3e79ad7e2cbf227eb9fd699838fb28b9fe988376c882ab8933d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4mkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8wxqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:21Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.119740 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4e7ae8a-acd5-4685-97b4-9f775d5c08d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80b7097545fca095740245eb7373a8ad9325bfde309725cf0b0db792bbd44df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5376dc11940017714443d1a4cf6c61f323eb35b41d1ce9b76d73137465f18c83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70c7310e5734c08f62e5732d069aed4eb9ccd5cc0e8ae546950871772f6a27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66bb948c15a1089cc29548df9da9f67b1f554d907a02f6076a0da38cabca9670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66bb948c15a1089cc29548df9da9f67b1f554d907a02f6076a0da38cabca9670\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:21Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.139075 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fa69354-6ea0-44dd-ad6b-7926364115f0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a5dbb64c2a6edf9d29acfce538f3747988606488c9c2dced7de40d230174a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be2ca58a1618a315c54c047c756d9878ba21af8846047cfc91c74fb23ddd9a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e398b423583958c7c6bce08282708e18bdaae5d3d492767dd1b397959e93ced4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543de8418ec56de4cab158f131bc2fb1421445a639bd4eeea321e60482e6175\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:21Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.156767 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d459b25-158a-4a2f-bff7-b7c643450c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc5172a4b40f579c4ae113095afcdc882c281d31c0ddf2f59f0593c0a8e1837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69b51611cd90f67fb36b564db5890aff54f52a1884d71d297f335d707ec8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:55:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtv8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:55:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5fvqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:56:21Z is after 2025-08-24T17:21:41Z" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.197130 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.197186 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.197199 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.197214 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.197229 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:21Z","lastTransitionTime":"2025-11-24T06:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.299273 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.299600 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.299611 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.299627 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.299637 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:21Z","lastTransitionTime":"2025-11-24T06:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.402883 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.402928 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.402937 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.402952 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.402985 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:21Z","lastTransitionTime":"2025-11-24T06:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.506570 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.506646 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.506667 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.506723 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.506744 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:21Z","lastTransitionTime":"2025-11-24T06:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.610366 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.610456 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.610480 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.610510 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.610529 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:21Z","lastTransitionTime":"2025-11-24T06:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.713832 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.713901 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.713923 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.713953 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.714022 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:21Z","lastTransitionTime":"2025-11-24T06:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.818055 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.818162 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.818189 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.818228 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.818255 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:21Z","lastTransitionTime":"2025-11-24T06:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.890957 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.891012 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:21 crc kubenswrapper[4809]: E1124 06:56:21.891122 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:21 crc kubenswrapper[4809]: E1124 06:56:21.891204 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.922494 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.922537 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.922549 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.922567 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:21 crc kubenswrapper[4809]: I1124 06:56:21.922581 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:21Z","lastTransitionTime":"2025-11-24T06:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.025395 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.025464 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.025475 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.025492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.025506 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:22Z","lastTransitionTime":"2025-11-24T06:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.128011 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.128072 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.128089 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.128109 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.128121 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:22Z","lastTransitionTime":"2025-11-24T06:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.231440 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.231736 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.231789 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.231820 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.231844 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:22Z","lastTransitionTime":"2025-11-24T06:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.335570 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.335646 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.335665 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.335690 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.335711 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:22Z","lastTransitionTime":"2025-11-24T06:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.439280 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.439350 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.439369 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.439397 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.439416 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:22Z","lastTransitionTime":"2025-11-24T06:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.541770 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.541814 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.541823 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.541837 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.541847 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:22Z","lastTransitionTime":"2025-11-24T06:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.645249 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.645315 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.645332 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.645355 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.645372 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:22Z","lastTransitionTime":"2025-11-24T06:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.748575 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.748640 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.748657 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.748681 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.748698 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:22Z","lastTransitionTime":"2025-11-24T06:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.851737 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.851791 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.851807 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.851832 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.851849 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:22Z","lastTransitionTime":"2025-11-24T06:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.890818 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.890870 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:22 crc kubenswrapper[4809]: E1124 06:56:22.891012 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:22 crc kubenswrapper[4809]: E1124 06:56:22.891146 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.954338 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.954404 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.954428 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.954462 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:22 crc kubenswrapper[4809]: I1124 06:56:22.954486 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:22Z","lastTransitionTime":"2025-11-24T06:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.057532 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.057575 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.057587 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.057606 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.057618 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:23Z","lastTransitionTime":"2025-11-24T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.160144 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.160175 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.160182 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.160194 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.160203 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:23Z","lastTransitionTime":"2025-11-24T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.262897 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.263035 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.263073 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.263102 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.263128 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:23Z","lastTransitionTime":"2025-11-24T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.366251 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.366291 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.366301 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.366337 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.366348 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:23Z","lastTransitionTime":"2025-11-24T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.471735 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.471805 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.471826 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.471855 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.471877 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:23Z","lastTransitionTime":"2025-11-24T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.575176 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.575234 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.575243 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.575261 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.575272 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:23Z","lastTransitionTime":"2025-11-24T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.678658 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.678736 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.678746 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.678770 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.678782 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:23Z","lastTransitionTime":"2025-11-24T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.780806 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.780882 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.780900 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.780917 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.780929 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:23Z","lastTransitionTime":"2025-11-24T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.883744 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.883786 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.883794 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.883808 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.883817 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:23Z","lastTransitionTime":"2025-11-24T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.891170 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.891193 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:23 crc kubenswrapper[4809]: E1124 06:56:23.891302 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:23 crc kubenswrapper[4809]: E1124 06:56:23.891381 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.911042 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.911202 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:23 crc kubenswrapper[4809]: E1124 06:56:23.911233 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:27.911208188 +0000 UTC m=+147.811799903 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.911274 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:23 crc kubenswrapper[4809]: E1124 06:56:23.911339 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:56:23 crc kubenswrapper[4809]: E1124 06:56:23.911373 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:56:23 crc kubenswrapper[4809]: E1124 06:56:23.911391 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:57:27.911374783 +0000 UTC m=+147.811966498 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:56:23 crc kubenswrapper[4809]: E1124 06:56:23.911410 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:57:27.911401034 +0000 UTC m=+147.811992749 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.987206 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.987269 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.987287 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.987328 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:23 crc kubenswrapper[4809]: I1124 06:56:23.987346 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:23Z","lastTransitionTime":"2025-11-24T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.013026 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.013121 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:24 crc kubenswrapper[4809]: E1124 06:56:24.013304 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:56:24 crc kubenswrapper[4809]: E1124 06:56:24.013354 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:56:24 crc kubenswrapper[4809]: E1124 06:56:24.013382 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:56:24 crc kubenswrapper[4809]: E1124 06:56:24.013492 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 06:57:28.013462838 +0000 UTC m=+147.914054593 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:56:24 crc kubenswrapper[4809]: E1124 06:56:24.013505 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:56:24 crc kubenswrapper[4809]: E1124 06:56:24.013558 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:56:24 crc kubenswrapper[4809]: E1124 06:56:24.013588 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:56:24 crc kubenswrapper[4809]: E1124 06:56:24.013690 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 06:57:28.013660503 +0000 UTC m=+147.914252248 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.089873 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.089938 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.089957 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.090013 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.090032 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:24Z","lastTransitionTime":"2025-11-24T06:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.192492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.192529 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.192538 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.192552 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.192562 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:24Z","lastTransitionTime":"2025-11-24T06:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.294781 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.294819 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.294828 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.294842 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.294852 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:24Z","lastTransitionTime":"2025-11-24T06:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.398281 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.398507 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.398525 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.398551 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.398567 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:24Z","lastTransitionTime":"2025-11-24T06:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.501764 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.501821 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.501838 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.501860 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.501877 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:24Z","lastTransitionTime":"2025-11-24T06:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.605388 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.605422 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.605432 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.605446 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.605456 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:24Z","lastTransitionTime":"2025-11-24T06:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.708657 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.708750 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.708775 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.708803 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.708824 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:24Z","lastTransitionTime":"2025-11-24T06:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.811136 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.811187 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.811197 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.811211 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.811221 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:24Z","lastTransitionTime":"2025-11-24T06:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.890535 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:24 crc kubenswrapper[4809]: E1124 06:56:24.890664 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.890717 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:24 crc kubenswrapper[4809]: E1124 06:56:24.890842 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.913003 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.913041 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.913049 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.913063 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:24 crc kubenswrapper[4809]: I1124 06:56:24.913072 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:24Z","lastTransitionTime":"2025-11-24T06:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.015782 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.015832 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.015846 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.015862 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.015872 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:25Z","lastTransitionTime":"2025-11-24T06:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.119349 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.119409 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.119426 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.119448 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.119467 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:25Z","lastTransitionTime":"2025-11-24T06:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.222697 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.222734 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.222742 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.222756 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.222765 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:25Z","lastTransitionTime":"2025-11-24T06:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.325923 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.325997 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.326009 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.326028 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.326041 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:25Z","lastTransitionTime":"2025-11-24T06:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.428533 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.428612 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.428638 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.428669 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.428692 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:25Z","lastTransitionTime":"2025-11-24T06:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.532613 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.532694 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.532719 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.532753 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.532779 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:25Z","lastTransitionTime":"2025-11-24T06:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.635706 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.635766 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.635783 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.635807 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.635823 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:25Z","lastTransitionTime":"2025-11-24T06:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.739265 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.739325 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.739343 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.739368 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.739384 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:25Z","lastTransitionTime":"2025-11-24T06:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.842897 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.842998 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.843018 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.843041 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.843059 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:25Z","lastTransitionTime":"2025-11-24T06:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.891138 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.891153 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:25 crc kubenswrapper[4809]: E1124 06:56:25.891309 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:25 crc kubenswrapper[4809]: E1124 06:56:25.891448 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.951063 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.951164 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.951326 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.951369 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:25 crc kubenswrapper[4809]: I1124 06:56:25.951408 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:25Z","lastTransitionTime":"2025-11-24T06:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.054405 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.054507 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.054527 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.054552 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.054569 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:26Z","lastTransitionTime":"2025-11-24T06:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.158113 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.158175 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.158192 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.158218 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.158239 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:26Z","lastTransitionTime":"2025-11-24T06:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.261957 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.262144 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.262212 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.262237 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.262255 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:26Z","lastTransitionTime":"2025-11-24T06:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.366435 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.366512 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.366530 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.366554 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.366571 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:26Z","lastTransitionTime":"2025-11-24T06:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.469568 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.469626 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.469645 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.469669 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.469687 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:26Z","lastTransitionTime":"2025-11-24T06:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.573326 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.573388 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.573407 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.573437 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.573459 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:26Z","lastTransitionTime":"2025-11-24T06:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.675622 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.675665 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.675674 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.675690 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.675700 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:26Z","lastTransitionTime":"2025-11-24T06:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.778695 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.778748 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.778763 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.778783 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.778798 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:26Z","lastTransitionTime":"2025-11-24T06:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.882068 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.882137 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.882157 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.882183 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.882203 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:26Z","lastTransitionTime":"2025-11-24T06:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.891338 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.891662 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:26 crc kubenswrapper[4809]: E1124 06:56:26.892228 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:26 crc kubenswrapper[4809]: E1124 06:56:26.892368 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.892466 4809 scope.go:117] "RemoveContainer" containerID="7f3517cb223a94e59385d72b6844b837a34cb285128c89db3ef6f259a4d505f3" Nov 24 06:56:26 crc kubenswrapper[4809]: E1124 06:56:26.892690 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-kpqhl_openshift-ovn-kubernetes(74622ad8-44b0-44a7-afbd-e19bb9635520)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.984729 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.984812 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.984851 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.984884 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:26 crc kubenswrapper[4809]: I1124 06:56:26.984906 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:26Z","lastTransitionTime":"2025-11-24T06:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.092197 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.092252 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.092269 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.092292 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.092309 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:27Z","lastTransitionTime":"2025-11-24T06:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.195616 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.195679 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.195704 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.195733 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.195755 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:27Z","lastTransitionTime":"2025-11-24T06:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.299386 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.299448 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.299466 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.299492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.299509 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:27Z","lastTransitionTime":"2025-11-24T06:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.402625 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.402688 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.402723 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.402743 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.402754 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:27Z","lastTransitionTime":"2025-11-24T06:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.505606 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.505663 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.505673 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.505704 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.505716 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:27Z","lastTransitionTime":"2025-11-24T06:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.608422 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.608476 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.608488 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.608502 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.608511 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:27Z","lastTransitionTime":"2025-11-24T06:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.710229 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.710268 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.710278 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.710309 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.710319 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:27Z","lastTransitionTime":"2025-11-24T06:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.813481 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.813550 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.813562 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.813583 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.813597 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:27Z","lastTransitionTime":"2025-11-24T06:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.890616 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.890658 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:27 crc kubenswrapper[4809]: E1124 06:56:27.890799 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:27 crc kubenswrapper[4809]: E1124 06:56:27.890937 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.916934 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.917015 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.917028 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.917045 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:27 crc kubenswrapper[4809]: I1124 06:56:27.917057 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:27Z","lastTransitionTime":"2025-11-24T06:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.020614 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.020694 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.020712 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.020736 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.020756 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:28Z","lastTransitionTime":"2025-11-24T06:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.127536 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.127772 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.128157 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.128650 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.128700 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:28Z","lastTransitionTime":"2025-11-24T06:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.231841 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.231878 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.231886 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.231916 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.231926 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:28Z","lastTransitionTime":"2025-11-24T06:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.335665 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.335713 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.335725 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.335741 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.335753 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:28Z","lastTransitionTime":"2025-11-24T06:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.437518 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.437574 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.437590 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.437612 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.437629 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:28Z","lastTransitionTime":"2025-11-24T06:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.521248 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.521286 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.521294 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.521308 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.521318 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:56:28Z","lastTransitionTime":"2025-11-24T06:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.577949 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-j2p2f"] Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.578736 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j2p2f" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.581376 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.581396 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.582188 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.583654 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.626110 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=14.626092616 podStartE2EDuration="14.626092616s" podCreationTimestamp="2025-11-24 06:56:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:56:28.598931817 +0000 UTC m=+88.499523532" watchObservedRunningTime="2025-11-24 06:56:28.626092616 +0000 UTC m=+88.526684331" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.640143 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-7cq8b" podStartSLOduration=67.640116854 podStartE2EDuration="1m7.640116854s" podCreationTimestamp="2025-11-24 06:55:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:56:28.639518115 +0000 UTC m=+88.540109830" watchObservedRunningTime="2025-11-24 06:56:28.640116854 +0000 UTC m=+88.540708579" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.666693 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/4b88d3c6-e529-40cc-abf6-721c495818de-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-j2p2f\" (UID: \"4b88d3c6-e529-40cc-abf6-721c495818de\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j2p2f" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.666737 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4b88d3c6-e529-40cc-abf6-721c495818de-service-ca\") pod \"cluster-version-operator-5c965bbfc6-j2p2f\" (UID: \"4b88d3c6-e529-40cc-abf6-721c495818de\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j2p2f" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.666796 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b88d3c6-e529-40cc-abf6-721c495818de-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-j2p2f\" (UID: \"4b88d3c6-e529-40cc-abf6-721c495818de\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j2p2f" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.666834 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b88d3c6-e529-40cc-abf6-721c495818de-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-j2p2f\" (UID: \"4b88d3c6-e529-40cc-abf6-721c495818de\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j2p2f" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.666868 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/4b88d3c6-e529-40cc-abf6-721c495818de-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-j2p2f\" (UID: \"4b88d3c6-e529-40cc-abf6-721c495818de\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j2p2f" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.670185 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=68.670156285 podStartE2EDuration="1m8.670156285s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:56:28.658209459 +0000 UTC m=+88.558801184" watchObservedRunningTime="2025-11-24 06:56:28.670156285 +0000 UTC m=+88.570748040" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.737495 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podStartSLOduration=67.737476921 podStartE2EDuration="1m7.737476921s" podCreationTimestamp="2025-11-24 06:55:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:56:28.737434089 +0000 UTC m=+88.638025784" watchObservedRunningTime="2025-11-24 06:56:28.737476921 +0000 UTC m=+88.638068626" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.737916 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-8p7tz" podStartSLOduration=68.737909353 podStartE2EDuration="1m8.737909353s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:56:28.726996387 +0000 UTC m=+88.627588092" watchObservedRunningTime="2025-11-24 06:56:28.737909353 +0000 UTC m=+88.638501058" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.750447 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-5xvl8" podStartSLOduration=67.750429057 podStartE2EDuration="1m7.750429057s" podCreationTimestamp="2025-11-24 06:55:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:56:28.750234162 +0000 UTC m=+88.650825887" watchObservedRunningTime="2025-11-24 06:56:28.750429057 +0000 UTC m=+88.651020762" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.767479 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/4b88d3c6-e529-40cc-abf6-721c495818de-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-j2p2f\" (UID: \"4b88d3c6-e529-40cc-abf6-721c495818de\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j2p2f" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.767514 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4b88d3c6-e529-40cc-abf6-721c495818de-service-ca\") pod \"cluster-version-operator-5c965bbfc6-j2p2f\" (UID: \"4b88d3c6-e529-40cc-abf6-721c495818de\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j2p2f" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.767543 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b88d3c6-e529-40cc-abf6-721c495818de-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-j2p2f\" (UID: \"4b88d3c6-e529-40cc-abf6-721c495818de\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j2p2f" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.767546 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/4b88d3c6-e529-40cc-abf6-721c495818de-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-j2p2f\" (UID: \"4b88d3c6-e529-40cc-abf6-721c495818de\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j2p2f" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.767565 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b88d3c6-e529-40cc-abf6-721c495818de-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-j2p2f\" (UID: \"4b88d3c6-e529-40cc-abf6-721c495818de\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j2p2f" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.767601 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/4b88d3c6-e529-40cc-abf6-721c495818de-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-j2p2f\" (UID: \"4b88d3c6-e529-40cc-abf6-721c495818de\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j2p2f" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.767644 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/4b88d3c6-e529-40cc-abf6-721c495818de-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-j2p2f\" (UID: \"4b88d3c6-e529-40cc-abf6-721c495818de\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j2p2f" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.768553 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4b88d3c6-e529-40cc-abf6-721c495818de-service-ca\") pod \"cluster-version-operator-5c965bbfc6-j2p2f\" (UID: \"4b88d3c6-e529-40cc-abf6-721c495818de\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j2p2f" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.779549 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b88d3c6-e529-40cc-abf6-721c495818de-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-j2p2f\" (UID: \"4b88d3c6-e529-40cc-abf6-721c495818de\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j2p2f" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.785750 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b88d3c6-e529-40cc-abf6-721c495818de-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-j2p2f\" (UID: \"4b88d3c6-e529-40cc-abf6-721c495818de\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j2p2f" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.790005 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=31.789985886 podStartE2EDuration="31.789985886s" podCreationTimestamp="2025-11-24 06:55:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:56:28.789405119 +0000 UTC m=+88.689996824" watchObservedRunningTime="2025-11-24 06:56:28.789985886 +0000 UTC m=+88.690577601" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.803510 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=68.803496319 podStartE2EDuration="1m8.803496319s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:56:28.80284274 +0000 UTC m=+88.703434465" watchObservedRunningTime="2025-11-24 06:56:28.803496319 +0000 UTC m=+88.704088024" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.847131 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-8wxqg" podStartSLOduration=67.847112465 podStartE2EDuration="1m7.847112465s" podCreationTimestamp="2025-11-24 06:55:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:56:28.846534408 +0000 UTC m=+88.747126113" watchObservedRunningTime="2025-11-24 06:56:28.847112465 +0000 UTC m=+88.747704170" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.890162 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.890162 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:28 crc kubenswrapper[4809]: E1124 06:56:28.890533 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:28 crc kubenswrapper[4809]: E1124 06:56:28.890584 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:28 crc kubenswrapper[4809]: I1124 06:56:28.894515 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j2p2f" Nov 24 06:56:28 crc kubenswrapper[4809]: W1124 06:56:28.907106 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b88d3c6_e529_40cc_abf6_721c495818de.slice/crio-25d794953d2a2d3c9966a746fbfb924d7b465cb38e9a58933c4e1cdb2eeea85b WatchSource:0}: Error finding container 25d794953d2a2d3c9966a746fbfb924d7b465cb38e9a58933c4e1cdb2eeea85b: Status 404 returned error can't find the container with id 25d794953d2a2d3c9966a746fbfb924d7b465cb38e9a58933c4e1cdb2eeea85b Nov 24 06:56:29 crc kubenswrapper[4809]: I1124 06:56:29.112761 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j2p2f" event={"ID":"4b88d3c6-e529-40cc-abf6-721c495818de","Type":"ContainerStarted","Data":"8317a6db5df1edff9a0f9a4eb63e41c1be3d07e64978b74fd2e035e11f093d4a"} Nov 24 06:56:29 crc kubenswrapper[4809]: I1124 06:56:29.112890 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j2p2f" event={"ID":"4b88d3c6-e529-40cc-abf6-721c495818de","Type":"ContainerStarted","Data":"25d794953d2a2d3c9966a746fbfb924d7b465cb38e9a58933c4e1cdb2eeea85b"} Nov 24 06:56:29 crc kubenswrapper[4809]: I1124 06:56:29.129526 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j2p2f" podStartSLOduration=69.129491577 podStartE2EDuration="1m9.129491577s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:56:29.12889263 +0000 UTC m=+89.029484375" watchObservedRunningTime="2025-11-24 06:56:29.129491577 +0000 UTC m=+89.030083332" Nov 24 06:56:29 crc kubenswrapper[4809]: I1124 06:56:29.130309 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5fvqr" podStartSLOduration=69.130296531 podStartE2EDuration="1m9.130296531s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:56:28.858859537 +0000 UTC m=+88.759451242" watchObservedRunningTime="2025-11-24 06:56:29.130296531 +0000 UTC m=+89.030888276" Nov 24 06:56:29 crc kubenswrapper[4809]: I1124 06:56:29.890482 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:29 crc kubenswrapper[4809]: I1124 06:56:29.890514 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:29 crc kubenswrapper[4809]: E1124 06:56:29.890689 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:29 crc kubenswrapper[4809]: E1124 06:56:29.890810 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:30 crc kubenswrapper[4809]: I1124 06:56:30.890467 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:30 crc kubenswrapper[4809]: I1124 06:56:30.890526 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:30 crc kubenswrapper[4809]: E1124 06:56:30.891882 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:30 crc kubenswrapper[4809]: E1124 06:56:30.892115 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:31 crc kubenswrapper[4809]: I1124 06:56:31.891245 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:31 crc kubenswrapper[4809]: I1124 06:56:31.891275 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:31 crc kubenswrapper[4809]: E1124 06:56:31.891400 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:31 crc kubenswrapper[4809]: E1124 06:56:31.891552 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:32 crc kubenswrapper[4809]: I1124 06:56:32.890850 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:32 crc kubenswrapper[4809]: I1124 06:56:32.890811 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:32 crc kubenswrapper[4809]: E1124 06:56:32.891224 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:32 crc kubenswrapper[4809]: E1124 06:56:32.891439 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:33 crc kubenswrapper[4809]: I1124 06:56:33.891053 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:33 crc kubenswrapper[4809]: I1124 06:56:33.891126 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:33 crc kubenswrapper[4809]: E1124 06:56:33.891295 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:33 crc kubenswrapper[4809]: E1124 06:56:33.891451 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:34 crc kubenswrapper[4809]: I1124 06:56:34.891274 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:34 crc kubenswrapper[4809]: I1124 06:56:34.891574 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:34 crc kubenswrapper[4809]: E1124 06:56:34.891712 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:34 crc kubenswrapper[4809]: E1124 06:56:34.891856 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:35 crc kubenswrapper[4809]: I1124 06:56:35.891078 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:35 crc kubenswrapper[4809]: I1124 06:56:35.891213 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:35 crc kubenswrapper[4809]: E1124 06:56:35.891252 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:35 crc kubenswrapper[4809]: E1124 06:56:35.891438 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:35 crc kubenswrapper[4809]: I1124 06:56:35.913059 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 24 06:56:36 crc kubenswrapper[4809]: I1124 06:56:36.890776 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:36 crc kubenswrapper[4809]: I1124 06:56:36.890841 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:36 crc kubenswrapper[4809]: E1124 06:56:36.891320 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:36 crc kubenswrapper[4809]: E1124 06:56:36.891398 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:37 crc kubenswrapper[4809]: I1124 06:56:37.890868 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:37 crc kubenswrapper[4809]: I1124 06:56:37.890930 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:37 crc kubenswrapper[4809]: E1124 06:56:37.891037 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:37 crc kubenswrapper[4809]: E1124 06:56:37.891137 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:38 crc kubenswrapper[4809]: I1124 06:56:38.469824 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs\") pod \"network-metrics-daemon-krmz8\" (UID: \"35b8615e-faa5-47b4-8c22-609d5fda9590\") " pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:38 crc kubenswrapper[4809]: E1124 06:56:38.470105 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:56:38 crc kubenswrapper[4809]: E1124 06:56:38.470211 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs podName:35b8615e-faa5-47b4-8c22-609d5fda9590 nodeName:}" failed. No retries permitted until 2025-11-24 06:57:42.470186815 +0000 UTC m=+162.370778560 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs") pod "network-metrics-daemon-krmz8" (UID: "35b8615e-faa5-47b4-8c22-609d5fda9590") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:56:38 crc kubenswrapper[4809]: I1124 06:56:38.890355 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:38 crc kubenswrapper[4809]: I1124 06:56:38.890371 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:38 crc kubenswrapper[4809]: E1124 06:56:38.890693 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:38 crc kubenswrapper[4809]: E1124 06:56:38.890547 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:39 crc kubenswrapper[4809]: I1124 06:56:39.890530 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:39 crc kubenswrapper[4809]: I1124 06:56:39.890624 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:39 crc kubenswrapper[4809]: E1124 06:56:39.890697 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:39 crc kubenswrapper[4809]: E1124 06:56:39.890838 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:40 crc kubenswrapper[4809]: I1124 06:56:40.890583 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:40 crc kubenswrapper[4809]: I1124 06:56:40.890695 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:40 crc kubenswrapper[4809]: E1124 06:56:40.890781 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:40 crc kubenswrapper[4809]: E1124 06:56:40.890946 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:40 crc kubenswrapper[4809]: I1124 06:56:40.892041 4809 scope.go:117] "RemoveContainer" containerID="7f3517cb223a94e59385d72b6844b837a34cb285128c89db3ef6f259a4d505f3" Nov 24 06:56:40 crc kubenswrapper[4809]: E1124 06:56:40.892283 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-kpqhl_openshift-ovn-kubernetes(74622ad8-44b0-44a7-afbd-e19bb9635520)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" Nov 24 06:56:41 crc kubenswrapper[4809]: I1124 06:56:41.890778 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:41 crc kubenswrapper[4809]: I1124 06:56:41.890823 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:41 crc kubenswrapper[4809]: E1124 06:56:41.891011 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:41 crc kubenswrapper[4809]: E1124 06:56:41.891128 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:42 crc kubenswrapper[4809]: I1124 06:56:42.891170 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:42 crc kubenswrapper[4809]: E1124 06:56:42.891544 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:42 crc kubenswrapper[4809]: I1124 06:56:42.891170 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:42 crc kubenswrapper[4809]: E1124 06:56:42.892363 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:43 crc kubenswrapper[4809]: I1124 06:56:43.890778 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:43 crc kubenswrapper[4809]: I1124 06:56:43.890813 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:43 crc kubenswrapper[4809]: E1124 06:56:43.890901 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:43 crc kubenswrapper[4809]: E1124 06:56:43.891005 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:44 crc kubenswrapper[4809]: I1124 06:56:44.890463 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:44 crc kubenswrapper[4809]: I1124 06:56:44.890521 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:44 crc kubenswrapper[4809]: E1124 06:56:44.890670 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:44 crc kubenswrapper[4809]: E1124 06:56:44.891067 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:45 crc kubenswrapper[4809]: I1124 06:56:45.890427 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:45 crc kubenswrapper[4809]: I1124 06:56:45.890462 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:45 crc kubenswrapper[4809]: E1124 06:56:45.890615 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:45 crc kubenswrapper[4809]: E1124 06:56:45.890715 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:47 crc kubenswrapper[4809]: I1124 06:56:46.890644 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:47 crc kubenswrapper[4809]: E1124 06:56:46.890760 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:47 crc kubenswrapper[4809]: I1124 06:56:46.891164 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:47 crc kubenswrapper[4809]: E1124 06:56:46.891617 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:47 crc kubenswrapper[4809]: I1124 06:56:47.890538 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:47 crc kubenswrapper[4809]: I1124 06:56:47.890566 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:47 crc kubenswrapper[4809]: E1124 06:56:47.890734 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:47 crc kubenswrapper[4809]: E1124 06:56:47.890836 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:48 crc kubenswrapper[4809]: I1124 06:56:48.890960 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:48 crc kubenswrapper[4809]: E1124 06:56:48.891484 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:48 crc kubenswrapper[4809]: I1124 06:56:48.891528 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:48 crc kubenswrapper[4809]: E1124 06:56:48.892013 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:49 crc kubenswrapper[4809]: I1124 06:56:49.890650 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:49 crc kubenswrapper[4809]: I1124 06:56:49.890688 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:49 crc kubenswrapper[4809]: E1124 06:56:49.890824 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:49 crc kubenswrapper[4809]: E1124 06:56:49.890949 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:50 crc kubenswrapper[4809]: I1124 06:56:50.890695 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:50 crc kubenswrapper[4809]: I1124 06:56:50.890723 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:50 crc kubenswrapper[4809]: E1124 06:56:50.892382 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:50 crc kubenswrapper[4809]: E1124 06:56:50.892518 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:51 crc kubenswrapper[4809]: I1124 06:56:51.890411 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:51 crc kubenswrapper[4809]: I1124 06:56:51.890446 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:51 crc kubenswrapper[4809]: E1124 06:56:51.890881 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:51 crc kubenswrapper[4809]: E1124 06:56:51.891350 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:52 crc kubenswrapper[4809]: I1124 06:56:52.890765 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:52 crc kubenswrapper[4809]: I1124 06:56:52.890793 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:52 crc kubenswrapper[4809]: E1124 06:56:52.896578 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:52 crc kubenswrapper[4809]: E1124 06:56:52.896769 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:52 crc kubenswrapper[4809]: I1124 06:56:52.897896 4809 scope.go:117] "RemoveContainer" containerID="7f3517cb223a94e59385d72b6844b837a34cb285128c89db3ef6f259a4d505f3" Nov 24 06:56:52 crc kubenswrapper[4809]: E1124 06:56:52.898271 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-kpqhl_openshift-ovn-kubernetes(74622ad8-44b0-44a7-afbd-e19bb9635520)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" Nov 24 06:56:53 crc kubenswrapper[4809]: I1124 06:56:53.890424 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:53 crc kubenswrapper[4809]: I1124 06:56:53.890488 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:53 crc kubenswrapper[4809]: E1124 06:56:53.890592 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:53 crc kubenswrapper[4809]: E1124 06:56:53.890712 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:54 crc kubenswrapper[4809]: I1124 06:56:54.891176 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:54 crc kubenswrapper[4809]: I1124 06:56:54.891200 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:54 crc kubenswrapper[4809]: E1124 06:56:54.891397 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:54 crc kubenswrapper[4809]: E1124 06:56:54.891441 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:55 crc kubenswrapper[4809]: I1124 06:56:55.890368 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:55 crc kubenswrapper[4809]: I1124 06:56:55.890459 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:55 crc kubenswrapper[4809]: E1124 06:56:55.890547 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:55 crc kubenswrapper[4809]: E1124 06:56:55.890586 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:56 crc kubenswrapper[4809]: I1124 06:56:56.206433 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5xvl8_3c932894-56b4-4303-8176-d12b12686b09/kube-multus/1.log" Nov 24 06:56:56 crc kubenswrapper[4809]: I1124 06:56:56.207326 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5xvl8_3c932894-56b4-4303-8176-d12b12686b09/kube-multus/0.log" Nov 24 06:56:56 crc kubenswrapper[4809]: I1124 06:56:56.207401 4809 generic.go:334] "Generic (PLEG): container finished" podID="3c932894-56b4-4303-8176-d12b12686b09" containerID="d68cb2e2c7982f3c55f7d9263588e648ba155a930efc0266051aaa628ccb9db7" exitCode=1 Nov 24 06:56:56 crc kubenswrapper[4809]: I1124 06:56:56.207460 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5xvl8" event={"ID":"3c932894-56b4-4303-8176-d12b12686b09","Type":"ContainerDied","Data":"d68cb2e2c7982f3c55f7d9263588e648ba155a930efc0266051aaa628ccb9db7"} Nov 24 06:56:56 crc kubenswrapper[4809]: I1124 06:56:56.207511 4809 scope.go:117] "RemoveContainer" containerID="0dab5a07698372f18ffbb7a2879de0363306b68beaa22833638fcda48b19ff95" Nov 24 06:56:56 crc kubenswrapper[4809]: I1124 06:56:56.208257 4809 scope.go:117] "RemoveContainer" containerID="d68cb2e2c7982f3c55f7d9263588e648ba155a930efc0266051aaa628ccb9db7" Nov 24 06:56:56 crc kubenswrapper[4809]: E1124 06:56:56.208606 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-5xvl8_openshift-multus(3c932894-56b4-4303-8176-d12b12686b09)\"" pod="openshift-multus/multus-5xvl8" podUID="3c932894-56b4-4303-8176-d12b12686b09" Nov 24 06:56:56 crc kubenswrapper[4809]: I1124 06:56:56.232432 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=21.232405132 podStartE2EDuration="21.232405132s" podCreationTimestamp="2025-11-24 06:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:56:40.935933574 +0000 UTC m=+100.836525289" watchObservedRunningTime="2025-11-24 06:56:56.232405132 +0000 UTC m=+116.132996877" Nov 24 06:56:56 crc kubenswrapper[4809]: I1124 06:56:56.891157 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:56 crc kubenswrapper[4809]: I1124 06:56:56.891172 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:56 crc kubenswrapper[4809]: E1124 06:56:56.891261 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:56 crc kubenswrapper[4809]: E1124 06:56:56.891406 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:57 crc kubenswrapper[4809]: I1124 06:56:57.213904 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5xvl8_3c932894-56b4-4303-8176-d12b12686b09/kube-multus/1.log" Nov 24 06:56:57 crc kubenswrapper[4809]: I1124 06:56:57.891036 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:57 crc kubenswrapper[4809]: I1124 06:56:57.891154 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:57 crc kubenswrapper[4809]: E1124 06:56:57.891281 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:57 crc kubenswrapper[4809]: E1124 06:56:57.891653 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:56:58 crc kubenswrapper[4809]: I1124 06:56:58.891062 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:56:58 crc kubenswrapper[4809]: E1124 06:56:58.891181 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:56:58 crc kubenswrapper[4809]: I1124 06:56:58.891074 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:56:58 crc kubenswrapper[4809]: E1124 06:56:58.891542 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:56:59 crc kubenswrapper[4809]: I1124 06:56:59.890202 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:56:59 crc kubenswrapper[4809]: I1124 06:56:59.890224 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:56:59 crc kubenswrapper[4809]: E1124 06:56:59.891374 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:56:59 crc kubenswrapper[4809]: E1124 06:56:59.891397 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:57:00 crc kubenswrapper[4809]: I1124 06:57:00.890246 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:57:00 crc kubenswrapper[4809]: E1124 06:57:00.892374 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:57:00 crc kubenswrapper[4809]: I1124 06:57:00.892449 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:57:00 crc kubenswrapper[4809]: E1124 06:57:00.892699 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:57:00 crc kubenswrapper[4809]: E1124 06:57:00.894167 4809 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 24 06:57:01 crc kubenswrapper[4809]: I1124 06:57:01.890805 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:57:01 crc kubenswrapper[4809]: I1124 06:57:01.890863 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:57:01 crc kubenswrapper[4809]: E1124 06:57:01.891002 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:57:01 crc kubenswrapper[4809]: E1124 06:57:01.891106 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:57:01 crc kubenswrapper[4809]: E1124 06:57:01.971717 4809 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 06:57:02 crc kubenswrapper[4809]: I1124 06:57:02.890489 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:57:02 crc kubenswrapper[4809]: I1124 06:57:02.890604 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:57:02 crc kubenswrapper[4809]: E1124 06:57:02.890690 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:57:02 crc kubenswrapper[4809]: E1124 06:57:02.890822 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:57:03 crc kubenswrapper[4809]: I1124 06:57:03.890608 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:57:03 crc kubenswrapper[4809]: I1124 06:57:03.890651 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:57:03 crc kubenswrapper[4809]: E1124 06:57:03.890735 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:57:03 crc kubenswrapper[4809]: E1124 06:57:03.890865 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:57:04 crc kubenswrapper[4809]: I1124 06:57:04.890230 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:57:04 crc kubenswrapper[4809]: I1124 06:57:04.890291 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:57:04 crc kubenswrapper[4809]: E1124 06:57:04.890446 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:57:04 crc kubenswrapper[4809]: E1124 06:57:04.890576 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:57:05 crc kubenswrapper[4809]: I1124 06:57:05.890868 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:57:05 crc kubenswrapper[4809]: I1124 06:57:05.890931 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:57:05 crc kubenswrapper[4809]: E1124 06:57:05.891161 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:57:05 crc kubenswrapper[4809]: E1124 06:57:05.891302 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:57:06 crc kubenswrapper[4809]: I1124 06:57:06.890621 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:57:06 crc kubenswrapper[4809]: E1124 06:57:06.890836 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:57:06 crc kubenswrapper[4809]: I1124 06:57:06.891010 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:57:06 crc kubenswrapper[4809]: E1124 06:57:06.891196 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:57:06 crc kubenswrapper[4809]: E1124 06:57:06.972986 4809 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 06:57:07 crc kubenswrapper[4809]: I1124 06:57:07.890903 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:57:07 crc kubenswrapper[4809]: I1124 06:57:07.891098 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:57:07 crc kubenswrapper[4809]: I1124 06:57:07.891298 4809 scope.go:117] "RemoveContainer" containerID="d68cb2e2c7982f3c55f7d9263588e648ba155a930efc0266051aaa628ccb9db7" Nov 24 06:57:07 crc kubenswrapper[4809]: E1124 06:57:07.892195 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:57:07 crc kubenswrapper[4809]: E1124 06:57:07.892599 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:57:07 crc kubenswrapper[4809]: I1124 06:57:07.892659 4809 scope.go:117] "RemoveContainer" containerID="7f3517cb223a94e59385d72b6844b837a34cb285128c89db3ef6f259a4d505f3" Nov 24 06:57:08 crc kubenswrapper[4809]: I1124 06:57:08.254053 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5xvl8_3c932894-56b4-4303-8176-d12b12686b09/kube-multus/1.log" Nov 24 06:57:08 crc kubenswrapper[4809]: I1124 06:57:08.254422 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5xvl8" event={"ID":"3c932894-56b4-4303-8176-d12b12686b09","Type":"ContainerStarted","Data":"4f0050e9681c436b5268f8f5d15039d7e02d064e44ac5fecf5a52b0ad3cf3a81"} Nov 24 06:57:08 crc kubenswrapper[4809]: I1124 06:57:08.259291 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kpqhl_74622ad8-44b0-44a7-afbd-e19bb9635520/ovnkube-controller/3.log" Nov 24 06:57:08 crc kubenswrapper[4809]: I1124 06:57:08.265067 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerStarted","Data":"aac5161cc51c14f1071e02e84f4cd9e1e7656f9a466ea5088b41c34d2687f929"} Nov 24 06:57:08 crc kubenswrapper[4809]: I1124 06:57:08.265615 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:57:08 crc kubenswrapper[4809]: I1124 06:57:08.801766 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" podStartSLOduration=107.801744686 podStartE2EDuration="1m47.801744686s" podCreationTimestamp="2025-11-24 06:55:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:08.325582276 +0000 UTC m=+128.226173991" watchObservedRunningTime="2025-11-24 06:57:08.801744686 +0000 UTC m=+128.702336401" Nov 24 06:57:08 crc kubenswrapper[4809]: I1124 06:57:08.803088 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-krmz8"] Nov 24 06:57:08 crc kubenswrapper[4809]: I1124 06:57:08.803206 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:57:08 crc kubenswrapper[4809]: E1124 06:57:08.803308 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:57:08 crc kubenswrapper[4809]: I1124 06:57:08.890247 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:57:08 crc kubenswrapper[4809]: E1124 06:57:08.890359 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:57:09 crc kubenswrapper[4809]: I1124 06:57:09.890995 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:57:09 crc kubenswrapper[4809]: E1124 06:57:09.891478 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:57:09 crc kubenswrapper[4809]: I1124 06:57:09.891848 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:57:09 crc kubenswrapper[4809]: E1124 06:57:09.892067 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:57:10 crc kubenswrapper[4809]: I1124 06:57:10.890682 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:57:10 crc kubenswrapper[4809]: I1124 06:57:10.890884 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:57:10 crc kubenswrapper[4809]: E1124 06:57:10.892721 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krmz8" podUID="35b8615e-faa5-47b4-8c22-609d5fda9590" Nov 24 06:57:10 crc kubenswrapper[4809]: E1124 06:57:10.892888 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:57:11 crc kubenswrapper[4809]: I1124 06:57:11.890891 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:57:11 crc kubenswrapper[4809]: I1124 06:57:11.890938 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:57:11 crc kubenswrapper[4809]: E1124 06:57:11.891029 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:57:11 crc kubenswrapper[4809]: E1124 06:57:11.891167 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:57:12 crc kubenswrapper[4809]: I1124 06:57:12.890594 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:57:12 crc kubenswrapper[4809]: I1124 06:57:12.890727 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:57:12 crc kubenswrapper[4809]: I1124 06:57:12.894536 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 24 06:57:12 crc kubenswrapper[4809]: I1124 06:57:12.895375 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 24 06:57:12 crc kubenswrapper[4809]: I1124 06:57:12.895395 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 24 06:57:12 crc kubenswrapper[4809]: I1124 06:57:12.895593 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 24 06:57:13 crc kubenswrapper[4809]: I1124 06:57:13.890452 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:57:13 crc kubenswrapper[4809]: I1124 06:57:13.891246 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:57:13 crc kubenswrapper[4809]: I1124 06:57:13.894324 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 24 06:57:13 crc kubenswrapper[4809]: I1124 06:57:13.899470 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 24 06:57:14 crc kubenswrapper[4809]: I1124 06:57:14.663321 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.325764 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.384020 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-56j57"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.387493 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-56j57" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.390297 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-92k7t"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.391620 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.394450 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.395259 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.396881 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.404841 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.404951 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.405531 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.406504 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.407027 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.408835 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.409151 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.414343 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.429300 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.429503 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.429699 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.430039 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-88wkc"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.430073 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.430855 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.431360 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.431395 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.431534 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.431671 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.432383 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.435486 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.436020 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-lkdj2"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.436419 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.436697 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-cj64v"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.437220 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cj64v" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.437560 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-lkdj2" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.437933 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.438620 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.441946 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.442328 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.442495 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.442608 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.446477 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-lxf68"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.447159 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-69dgx"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.447358 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.447574 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.447628 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.447851 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.451005 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rffdv"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.451473 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rffdv" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.452526 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-lxf68" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.457942 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.458183 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.458907 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.459152 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.459379 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.459584 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.460364 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.460431 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.460650 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.462419 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-q8lpg"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.462934 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q8lpg" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.464998 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rpd99"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.465239 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nl2x7"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.465484 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nl2x7" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.466211 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-rpd99" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.484036 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.484566 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.484598 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.484863 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.484992 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.485103 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.485216 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.485392 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.485471 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.485553 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.485629 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.485715 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.485784 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.485866 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.486079 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.486088 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.486162 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.486213 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.486305 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.486340 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.486456 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.486583 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.486659 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.486696 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.494485 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.494762 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.504822 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.505303 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.505660 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.505943 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.506935 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.507251 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.508244 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.511182 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.511372 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.511493 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.511614 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.512374 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.512532 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.512678 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.513725 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.515534 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.515681 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.515802 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.515821 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.515887 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-8r5vw"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.515930 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.516053 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.516151 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.516328 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.516455 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.517340 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-r7k56"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.517718 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.518311 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-r7k56" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.519556 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.521326 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.537206 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-54rmc"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.537726 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/25780238-d742-4461-af80-a4808c4dd40b-machine-approver-tls\") pod \"machine-approver-56656f9798-cj64v\" (UID: \"25780238-d742-4461-af80-a4808c4dd40b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cj64v" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.537773 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj2x5\" (UniqueName: \"kubernetes.io/projected/dd0770b5-d9e3-48ae-9ce0-58bcd6a47207-kube-api-access-dj2x5\") pod \"openshift-controller-manager-operator-756b6f6bc6-rffdv\" (UID: \"dd0770b5-d9e3-48ae-9ce0-58bcd6a47207\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rffdv" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.537798 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/66ef6324-33df-4e2a-92dd-36a3f5f7995d-audit\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.537815 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/eca6947b-dd5e-449d-a8c5-05277c556bd1-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-56j57\" (UID: \"eca6947b-dd5e-449d-a8c5-05277c556bd1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-56j57" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.537828 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f3ac0673-5554-4ac6-a16e-b08631fdbf94-client-ca\") pod \"controller-manager-879f6c89f-92k7t\" (UID: \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.537842 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66ef6324-33df-4e2a-92dd-36a3f5f7995d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.537858 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-serving-cert\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.537873 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f1b5104-f6eb-4e51-9339-e63f4429f02c-trusted-ca\") pod \"console-operator-58897d9998-rpd99\" (UID: \"8f1b5104-f6eb-4e51-9339-e63f4429f02c\") " pod="openshift-console-operator/console-operator-58897d9998-rpd99" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.537891 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.537956 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.537983 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhczn\" (UniqueName: \"kubernetes.io/projected/66ef6324-33df-4e2a-92dd-36a3f5f7995d-kube-api-access-dhczn\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.537998 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8frv\" (UniqueName: \"kubernetes.io/projected/8f1b5104-f6eb-4e51-9339-e63f4429f02c-kube-api-access-z8frv\") pod \"console-operator-58897d9998-rpd99\" (UID: \"8f1b5104-f6eb-4e51-9339-e63f4429f02c\") " pod="openshift-console-operator/console-operator-58897d9998-rpd99" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538012 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-54rmc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.537732 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538082 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538122 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538012 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-config\") pod \"route-controller-manager-6576b87f9c-p86rl\" (UID: \"358682a8-6b18-49ee-bd7d-d9eab2b22ee0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538238 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538256 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3ac0673-5554-4ac6-a16e-b08631fdbf94-serving-cert\") pod \"controller-manager-879f6c89f-92k7t\" (UID: \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538277 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxqtp\" (UniqueName: \"kubernetes.io/projected/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-kube-api-access-sxqtp\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538299 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-audit-dir\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538339 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25780238-d742-4461-af80-a4808c4dd40b-config\") pod \"machine-approver-56656f9798-cj64v\" (UID: \"25780238-d742-4461-af80-a4808c4dd40b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cj64v" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538358 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvkg4\" (UniqueName: \"kubernetes.io/projected/25780238-d742-4461-af80-a4808c4dd40b-kube-api-access-xvkg4\") pod \"machine-approver-56656f9798-cj64v\" (UID: \"25780238-d742-4461-af80-a4808c4dd40b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cj64v" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538372 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/66ef6324-33df-4e2a-92dd-36a3f5f7995d-etcd-serving-ca\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538389 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/25780238-d742-4461-af80-a4808c4dd40b-auth-proxy-config\") pod \"machine-approver-56656f9798-cj64v\" (UID: \"25780238-d742-4461-af80-a4808c4dd40b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cj64v" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538425 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538440 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538525 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538568 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538587 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f1b5104-f6eb-4e51-9339-e63f4429f02c-config\") pod \"console-operator-58897d9998-rpd99\" (UID: \"8f1b5104-f6eb-4e51-9339-e63f4429f02c\") " pod="openshift-console-operator/console-operator-58897d9998-rpd99" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538604 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe7c3c56-9f9c-4120-9a89-f840aff6aac8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nl2x7\" (UID: \"fe7c3c56-9f9c-4120-9a89-f840aff6aac8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nl2x7" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538623 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538641 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538656 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjvxr\" (UniqueName: \"kubernetes.io/projected/1bace469-be4b-48d6-9152-d65a4b762f5d-kube-api-access-fjvxr\") pod \"ingress-operator-5b745b69d9-q8lpg\" (UID: \"1bace469-be4b-48d6-9152-d65a4b762f5d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q8lpg" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538673 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-etcd-client\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538687 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe7c3c56-9f9c-4120-9a89-f840aff6aac8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nl2x7\" (UID: \"fe7c3c56-9f9c-4120-9a89-f840aff6aac8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nl2x7" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538702 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/df85fd23-9a76-41a3-b17d-7d06b5bfda87-etcd-ca\") pod \"etcd-operator-b45778765-lxf68\" (UID: \"df85fd23-9a76-41a3-b17d-7d06b5bfda87\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lxf68" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538719 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538736 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f1b5104-f6eb-4e51-9339-e63f4429f02c-serving-cert\") pod \"console-operator-58897d9998-rpd99\" (UID: \"8f1b5104-f6eb-4e51-9339-e63f4429f02c\") " pod="openshift-console-operator/console-operator-58897d9998-rpd99" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538752 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538766 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1bace469-be4b-48d6-9152-d65a4b762f5d-metrics-tls\") pod \"ingress-operator-5b745b69d9-q8lpg\" (UID: \"1bace469-be4b-48d6-9152-d65a4b762f5d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q8lpg" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538829 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1bace469-be4b-48d6-9152-d65a4b762f5d-trusted-ca\") pod \"ingress-operator-5b745b69d9-q8lpg\" (UID: \"1bace469-be4b-48d6-9152-d65a4b762f5d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q8lpg" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538464 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.539087 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mmdhq"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.539436 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-mmdhq" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.538859 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.539741 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr6gp\" (UniqueName: \"kubernetes.io/projected/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-kube-api-access-wr6gp\") pod \"route-controller-manager-6576b87f9c-p86rl\" (UID: \"358682a8-6b18-49ee-bd7d-d9eab2b22ee0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.539769 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/df85fd23-9a76-41a3-b17d-7d06b5bfda87-etcd-client\") pod \"etcd-operator-b45778765-lxf68\" (UID: \"df85fd23-9a76-41a3-b17d-7d06b5bfda87\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lxf68" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.539785 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/df85fd23-9a76-41a3-b17d-7d06b5bfda87-etcd-service-ca\") pod \"etcd-operator-b45778765-lxf68\" (UID: \"df85fd23-9a76-41a3-b17d-7d06b5bfda87\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lxf68" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.539802 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/66ef6324-33df-4e2a-92dd-36a3f5f7995d-etcd-client\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.539817 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/66ef6324-33df-4e2a-92dd-36a3f5f7995d-image-import-ca\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.539866 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7vgf\" (UniqueName: \"kubernetes.io/projected/eca6947b-dd5e-449d-a8c5-05277c556bd1-kube-api-access-l7vgf\") pod \"machine-api-operator-5694c8668f-56j57\" (UID: \"eca6947b-dd5e-449d-a8c5-05277c556bd1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-56j57" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.539892 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df85fd23-9a76-41a3-b17d-7d06b5bfda87-config\") pod \"etcd-operator-b45778765-lxf68\" (UID: \"df85fd23-9a76-41a3-b17d-7d06b5bfda87\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lxf68" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.539958 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df85fd23-9a76-41a3-b17d-7d06b5bfda87-serving-cert\") pod \"etcd-operator-b45778765-lxf68\" (UID: \"df85fd23-9a76-41a3-b17d-7d06b5bfda87\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lxf68" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.539991 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/66ef6324-33df-4e2a-92dd-36a3f5f7995d-encryption-config\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540041 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r4sq\" (UniqueName: \"kubernetes.io/projected/f3ac0673-5554-4ac6-a16e-b08631fdbf94-kube-api-access-4r4sq\") pod \"controller-manager-879f6c89f-92k7t\" (UID: \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540071 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1bace469-be4b-48d6-9152-d65a4b762f5d-bound-sa-token\") pod \"ingress-operator-5b745b69d9-q8lpg\" (UID: \"1bace469-be4b-48d6-9152-d65a4b762f5d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q8lpg" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540088 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbmcq\" (UniqueName: \"kubernetes.io/projected/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-kube-api-access-rbmcq\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540123 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-audit-policies\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540140 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66ef6324-33df-4e2a-92dd-36a3f5f7995d-serving-cert\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540199 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nh4b\" (UniqueName: \"kubernetes.io/projected/fe7c3c56-9f9c-4120-9a89-f840aff6aac8-kube-api-access-8nh4b\") pod \"kube-storage-version-migrator-operator-b67b599dd-nl2x7\" (UID: \"fe7c3c56-9f9c-4120-9a89-f840aff6aac8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nl2x7" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540220 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3ac0673-5554-4ac6-a16e-b08631fdbf94-config\") pod \"controller-manager-879f6c89f-92k7t\" (UID: \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540238 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4mvw\" (UniqueName: \"kubernetes.io/projected/df85fd23-9a76-41a3-b17d-7d06b5bfda87-kube-api-access-m4mvw\") pod \"etcd-operator-b45778765-lxf68\" (UID: \"df85fd23-9a76-41a3-b17d-7d06b5bfda87\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lxf68" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540291 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-client-ca\") pod \"route-controller-manager-6576b87f9c-p86rl\" (UID: \"358682a8-6b18-49ee-bd7d-d9eab2b22ee0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540317 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540335 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f3ac0673-5554-4ac6-a16e-b08631fdbf94-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-92k7t\" (UID: \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540362 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66ef6324-33df-4e2a-92dd-36a3f5f7995d-audit-dir\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540378 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmwcj\" (UniqueName: \"kubernetes.io/projected/b6237cc7-ba93-45a6-b560-d7916abf71f7-kube-api-access-mmwcj\") pod \"downloads-7954f5f757-lkdj2\" (UID: \"b6237cc7-ba93-45a6-b560-d7916abf71f7\") " pod="openshift-console/downloads-7954f5f757-lkdj2" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540394 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-serving-cert\") pod \"route-controller-manager-6576b87f9c-p86rl\" (UID: \"358682a8-6b18-49ee-bd7d-d9eab2b22ee0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540435 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540452 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/eca6947b-dd5e-449d-a8c5-05277c556bd1-images\") pod \"machine-api-operator-5694c8668f-56j57\" (UID: \"eca6947b-dd5e-449d-a8c5-05277c556bd1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-56j57" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540467 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd0770b5-d9e3-48ae-9ce0-58bcd6a47207-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-rffdv\" (UID: \"dd0770b5-d9e3-48ae-9ce0-58bcd6a47207\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rffdv" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540490 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/66ef6324-33df-4e2a-92dd-36a3f5f7995d-node-pullsecrets\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540503 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd0770b5-d9e3-48ae-9ce0-58bcd6a47207-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-rffdv\" (UID: \"dd0770b5-d9e3-48ae-9ce0-58bcd6a47207\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rffdv" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540655 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eca6947b-dd5e-449d-a8c5-05277c556bd1-config\") pod \"machine-api-operator-5694c8668f-56j57\" (UID: \"eca6947b-dd5e-449d-a8c5-05277c556bd1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-56j57" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540696 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540712 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-audit-policies\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540729 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540744 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-audit-dir\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540758 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66ef6324-33df-4e2a-92dd-36a3f5f7995d-config\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.540804 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-encryption-config\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.541928 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.542044 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.542114 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.544233 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.544421 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zhgn2"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.545105 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zhgn2" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.545784 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m9m27"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.546408 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m9m27" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.547624 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.549043 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.551242 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.551816 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ggzn"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.552438 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ggzn" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.552602 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xq929"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.553081 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xq929" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.555021 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xfdvj"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.555520 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xfdvj" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.555518 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.555872 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.557486 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6jp2"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.557890 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prdvb"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.558263 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prdvb" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.558447 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6jp2" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.560789 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.562325 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-7s9hl"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.562875 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7s9hl" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.565155 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-56j57"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.566040 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-72wvs"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.567027 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-72wvs" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.569020 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q9xpc"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.569806 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-q9xpc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.571105 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pmsgp"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.571597 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pmsgp" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.573545 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-59cgv"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.574140 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.575417 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.577231 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mqs58"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.579827 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mqs58" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.585790 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-lkdj2"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.587253 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.587705 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-92k7t"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.588810 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jzwg5"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.589583 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzwg5" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.590767 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.592314 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-q8lpg"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.592933 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-lxf68"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.593959 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mmdhq"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.596757 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.597747 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-8r5vw"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.597807 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-69dgx"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.597878 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xq929"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.598879 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-72wvs"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.600217 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6jp2"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.601129 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xfdvj"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.602397 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-z2dgw"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.602883 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pmsgp"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.602958 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-z2dgw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.605995 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5w6mp"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.606754 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.607028 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5w6mp" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.607759 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.610086 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-jx6qd"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.610497 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-jx6qd" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.611925 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-tk28k"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.612288 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pvrfw"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.612561 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.613164 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-tk28k" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.613177 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-pvrfw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.613435 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399445-bkt2p"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.614156 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-bkt2p" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.614430 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-svxc5"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.614910 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.615468 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-88wkc"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.616728 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nl2x7"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.618010 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zhgn2"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.619119 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ggzn"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.620051 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prdvb"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.621223 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rffdv"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.622201 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q9xpc"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.623122 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jzwg5"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.624962 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m9m27"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.625998 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rpd99"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.627019 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-54rmc"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.627869 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-gzp8q"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.629401 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-wlshf"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.629601 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.634853 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.638860 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-7s9hl"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.638893 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mqs58"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.639041 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-wlshf" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.639419 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-z2dgw"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.641760 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.641790 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr6gp\" (UniqueName: \"kubernetes.io/projected/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-kube-api-access-wr6gp\") pod \"route-controller-manager-6576b87f9c-p86rl\" (UID: \"358682a8-6b18-49ee-bd7d-d9eab2b22ee0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.641814 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdr48\" (UniqueName: \"kubernetes.io/projected/9eab3672-3dd0-4af1-b96d-e8c5e99bb840-kube-api-access-hdr48\") pod \"router-default-5444994796-r7k56\" (UID: \"9eab3672-3dd0-4af1-b96d-e8c5e99bb840\") " pod="openshift-ingress/router-default-5444994796-r7k56" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.641837 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmvs7\" (UniqueName: \"kubernetes.io/projected/8b246b95-be2e-4c90-8cf0-f92c612bfbc1-kube-api-access-lmvs7\") pod \"olm-operator-6b444d44fb-pmsgp\" (UID: \"8b246b95-be2e-4c90-8cf0-f92c612bfbc1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pmsgp" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.641854 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/df85fd23-9a76-41a3-b17d-7d06b5bfda87-etcd-client\") pod \"etcd-operator-b45778765-lxf68\" (UID: \"df85fd23-9a76-41a3-b17d-7d06b5bfda87\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lxf68" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.641870 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/df85fd23-9a76-41a3-b17d-7d06b5bfda87-etcd-service-ca\") pod \"etcd-operator-b45778765-lxf68\" (UID: \"df85fd23-9a76-41a3-b17d-7d06b5bfda87\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lxf68" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.641887 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/66ef6324-33df-4e2a-92dd-36a3f5f7995d-etcd-client\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.641901 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/66ef6324-33df-4e2a-92dd-36a3f5f7995d-image-import-ca\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.641916 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4-config\") pod \"authentication-operator-69f744f599-mmdhq\" (UID: \"8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmdhq" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.641941 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7vgf\" (UniqueName: \"kubernetes.io/projected/eca6947b-dd5e-449d-a8c5-05277c556bd1-kube-api-access-l7vgf\") pod \"machine-api-operator-5694c8668f-56j57\" (UID: \"eca6947b-dd5e-449d-a8c5-05277c556bd1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-56j57" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.641996 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df85fd23-9a76-41a3-b17d-7d06b5bfda87-config\") pod \"etcd-operator-b45778765-lxf68\" (UID: \"df85fd23-9a76-41a3-b17d-7d06b5bfda87\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lxf68" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642015 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtrcx\" (UniqueName: \"kubernetes.io/projected/c3b64526-f2a8-4141-b04c-65b65be09ccc-kube-api-access-xtrcx\") pod \"openshift-config-operator-7777fb866f-zhgn2\" (UID: \"c3b64526-f2a8-4141-b04c-65b65be09ccc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zhgn2" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642031 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8b246b95-be2e-4c90-8cf0-f92c612bfbc1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-pmsgp\" (UID: \"8b246b95-be2e-4c90-8cf0-f92c612bfbc1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pmsgp" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642048 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r4sq\" (UniqueName: \"kubernetes.io/projected/f3ac0673-5554-4ac6-a16e-b08631fdbf94-kube-api-access-4r4sq\") pod \"controller-manager-879f6c89f-92k7t\" (UID: \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642064 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1bace469-be4b-48d6-9152-d65a4b762f5d-bound-sa-token\") pod \"ingress-operator-5b745b69d9-q8lpg\" (UID: \"1bace469-be4b-48d6-9152-d65a4b762f5d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q8lpg" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642079 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df85fd23-9a76-41a3-b17d-7d06b5bfda87-serving-cert\") pod \"etcd-operator-b45778765-lxf68\" (UID: \"df85fd23-9a76-41a3-b17d-7d06b5bfda87\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lxf68" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642094 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/66ef6324-33df-4e2a-92dd-36a3f5f7995d-encryption-config\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642123 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3e14d56e-43b3-40ec-809f-371e8ab2f628-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-xfdvj\" (UID: \"3e14d56e-43b3-40ec-809f-371e8ab2f628\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xfdvj" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642150 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-audit-policies\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642172 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbmcq\" (UniqueName: \"kubernetes.io/projected/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-kube-api-access-rbmcq\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642197 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nh4b\" (UniqueName: \"kubernetes.io/projected/fe7c3c56-9f9c-4120-9a89-f840aff6aac8-kube-api-access-8nh4b\") pod \"kube-storage-version-migrator-operator-b67b599dd-nl2x7\" (UID: \"fe7c3c56-9f9c-4120-9a89-f840aff6aac8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nl2x7" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642219 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3ac0673-5554-4ac6-a16e-b08631fdbf94-config\") pod \"controller-manager-879f6c89f-92k7t\" (UID: \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642243 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4mvw\" (UniqueName: \"kubernetes.io/projected/df85fd23-9a76-41a3-b17d-7d06b5bfda87-kube-api-access-m4mvw\") pod \"etcd-operator-b45778765-lxf68\" (UID: \"df85fd23-9a76-41a3-b17d-7d06b5bfda87\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lxf68" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642264 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66ef6324-33df-4e2a-92dd-36a3f5f7995d-serving-cert\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642293 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fc65\" (UniqueName: \"kubernetes.io/projected/96284acd-e6e2-467c-8c70-005ca153dda3-kube-api-access-2fc65\") pod \"migrator-59844c95c7-54rmc\" (UID: \"96284acd-e6e2-467c-8c70-005ca153dda3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-54rmc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642298 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642322 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-client-ca\") pod \"route-controller-manager-6576b87f9c-p86rl\" (UID: \"358682a8-6b18-49ee-bd7d-d9eab2b22ee0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642342 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642358 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f3ac0673-5554-4ac6-a16e-b08631fdbf94-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-92k7t\" (UID: \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642375 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-service-ca\") pod \"console-f9d7485db-8r5vw\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642392 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4-serving-cert\") pod \"authentication-operator-69f744f599-mmdhq\" (UID: \"8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmdhq" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642411 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66ef6324-33df-4e2a-92dd-36a3f5f7995d-audit-dir\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642432 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r755l\" (UniqueName: \"kubernetes.io/projected/f1a1cfe0-beab-41d4-a478-668f5e7386f9-kube-api-access-r755l\") pod \"catalog-operator-68c6474976-72wvs\" (UID: \"f1a1cfe0-beab-41d4-a478-668f5e7386f9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-72wvs" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642451 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642469 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmwcj\" (UniqueName: \"kubernetes.io/projected/b6237cc7-ba93-45a6-b560-d7916abf71f7-kube-api-access-mmwcj\") pod \"downloads-7954f5f757-lkdj2\" (UID: \"b6237cc7-ba93-45a6-b560-d7916abf71f7\") " pod="openshift-console/downloads-7954f5f757-lkdj2" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642488 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-serving-cert\") pod \"route-controller-manager-6576b87f9c-p86rl\" (UID: \"358682a8-6b18-49ee-bd7d-d9eab2b22ee0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642505 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/7bf55845-293a-497e-87f1-eb5c36670105-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-mqs58\" (UID: \"7bf55845-293a-497e-87f1-eb5c36670105\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mqs58" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642521 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmffc\" (UniqueName: \"kubernetes.io/projected/cf7f6885-ff5e-44df-ac63-f00fa3ef63a6-kube-api-access-qmffc\") pod \"openshift-apiserver-operator-796bbdcf4f-m9m27\" (UID: \"cf7f6885-ff5e-44df-ac63-f00fa3ef63a6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m9m27" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642539 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd0770b5-d9e3-48ae-9ce0-58bcd6a47207-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-rffdv\" (UID: \"dd0770b5-d9e3-48ae-9ce0-58bcd6a47207\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rffdv" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642558 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/66ef6324-33df-4e2a-92dd-36a3f5f7995d-node-pullsecrets\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642574 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfjgk\" (UniqueName: \"kubernetes.io/projected/4e3c9b7d-5fa3-4f70-babe-522cb8dc7443-kube-api-access-rfjgk\") pod \"cluster-image-registry-operator-dc59b4c8b-q6jp2\" (UID: \"4e3c9b7d-5fa3-4f70-babe-522cb8dc7443\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6jp2" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642593 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/eca6947b-dd5e-449d-a8c5-05277c556bd1-images\") pod \"machine-api-operator-5694c8668f-56j57\" (UID: \"eca6947b-dd5e-449d-a8c5-05277c556bd1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-56j57" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642617 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eca6947b-dd5e-449d-a8c5-05277c556bd1-config\") pod \"machine-api-operator-5694c8668f-56j57\" (UID: \"eca6947b-dd5e-449d-a8c5-05277c556bd1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-56j57" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642634 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642653 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-audit-policies\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642670 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd0770b5-d9e3-48ae-9ce0-58bcd6a47207-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-rffdv\" (UID: \"dd0770b5-d9e3-48ae-9ce0-58bcd6a47207\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rffdv" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642684 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/4e3c9b7d-5fa3-4f70-babe-522cb8dc7443-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-q6jp2\" (UID: \"4e3c9b7d-5fa3-4f70-babe-522cb8dc7443\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6jp2" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642702 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-audit-dir\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642718 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66ef6324-33df-4e2a-92dd-36a3f5f7995d-config\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642732 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f1a1cfe0-beab-41d4-a478-668f5e7386f9-profile-collector-cert\") pod \"catalog-operator-68c6474976-72wvs\" (UID: \"f1a1cfe0-beab-41d4-a478-668f5e7386f9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-72wvs" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642749 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq84x\" (UniqueName: \"kubernetes.io/projected/7bf55845-293a-497e-87f1-eb5c36670105-kube-api-access-hq84x\") pod \"cluster-samples-operator-665b6dd947-mqs58\" (UID: \"7bf55845-293a-497e-87f1-eb5c36670105\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mqs58" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642795 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642811 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-encryption-config\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642827 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-console-config\") pod \"console-f9d7485db-8r5vw\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642843 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntbxx\" (UniqueName: \"kubernetes.io/projected/3e14d56e-43b3-40ec-809f-371e8ab2f628-kube-api-access-ntbxx\") pod \"control-plane-machine-set-operator-78cbb6b69f-xfdvj\" (UID: \"3e14d56e-43b3-40ec-809f-371e8ab2f628\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xfdvj" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642860 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/25780238-d742-4461-af80-a4808c4dd40b-machine-approver-tls\") pod \"machine-approver-56656f9798-cj64v\" (UID: \"25780238-d742-4461-af80-a4808c4dd40b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cj64v" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642876 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj2x5\" (UniqueName: \"kubernetes.io/projected/dd0770b5-d9e3-48ae-9ce0-58bcd6a47207-kube-api-access-dj2x5\") pod \"openshift-controller-manager-operator-756b6f6bc6-rffdv\" (UID: \"dd0770b5-d9e3-48ae-9ce0-58bcd6a47207\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rffdv" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642892 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/66ef6324-33df-4e2a-92dd-36a3f5f7995d-audit\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642910 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-console-oauth-config\") pod \"console-f9d7485db-8r5vw\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.642946 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f3ac0673-5554-4ac6-a16e-b08631fdbf94-client-ca\") pod \"controller-manager-879f6c89f-92k7t\" (UID: \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643002 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66ef6324-33df-4e2a-92dd-36a3f5f7995d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643022 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-trusted-ca-bundle\") pod \"console-f9d7485db-8r5vw\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643037 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4e3c9b7d-5fa3-4f70-babe-522cb8dc7443-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-q6jp2\" (UID: \"4e3c9b7d-5fa3-4f70-babe-522cb8dc7443\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6jp2" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643054 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/eca6947b-dd5e-449d-a8c5-05277c556bd1-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-56j57\" (UID: \"eca6947b-dd5e-449d-a8c5-05277c556bd1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-56j57" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643070 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-serving-cert\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643087 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4-service-ca-bundle\") pod \"authentication-operator-69f744f599-mmdhq\" (UID: \"8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmdhq" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643106 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f1b5104-f6eb-4e51-9339-e63f4429f02c-trusted-ca\") pod \"console-operator-58897d9998-rpd99\" (UID: \"8f1b5104-f6eb-4e51-9339-e63f4429f02c\") " pod="openshift-console-operator/console-operator-58897d9998-rpd99" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643122 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643138 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhczn\" (UniqueName: \"kubernetes.io/projected/66ef6324-33df-4e2a-92dd-36a3f5f7995d-kube-api-access-dhczn\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643162 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9eab3672-3dd0-4af1-b96d-e8c5e99bb840-default-certificate\") pod \"router-default-5444994796-r7k56\" (UID: \"9eab3672-3dd0-4af1-b96d-e8c5e99bb840\") " pod="openshift-ingress/router-default-5444994796-r7k56" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643179 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-oauth-serving-cert\") pod \"console-f9d7485db-8r5vw\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643195 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4e3c9b7d-5fa3-4f70-babe-522cb8dc7443-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-q6jp2\" (UID: \"4e3c9b7d-5fa3-4f70-babe-522cb8dc7443\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6jp2" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643212 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8b246b95-be2e-4c90-8cf0-f92c612bfbc1-srv-cert\") pod \"olm-operator-6b444d44fb-pmsgp\" (UID: \"8b246b95-be2e-4c90-8cf0-f92c612bfbc1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pmsgp" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643231 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643249 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643251 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8frv\" (UniqueName: \"kubernetes.io/projected/8f1b5104-f6eb-4e51-9339-e63f4429f02c-kube-api-access-z8frv\") pod \"console-operator-58897d9998-rpd99\" (UID: \"8f1b5104-f6eb-4e51-9339-e63f4429f02c\") " pod="openshift-console-operator/console-operator-58897d9998-rpd99" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643291 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-config\") pod \"route-controller-manager-6576b87f9c-p86rl\" (UID: \"358682a8-6b18-49ee-bd7d-d9eab2b22ee0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643313 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3ac0673-5554-4ac6-a16e-b08631fdbf94-serving-cert\") pod \"controller-manager-879f6c89f-92k7t\" (UID: \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643338 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxqtp\" (UniqueName: \"kubernetes.io/projected/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-kube-api-access-sxqtp\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643361 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9eab3672-3dd0-4af1-b96d-e8c5e99bb840-metrics-certs\") pod \"router-default-5444994796-r7k56\" (UID: \"9eab3672-3dd0-4af1-b96d-e8c5e99bb840\") " pod="openshift-ingress/router-default-5444994796-r7k56" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643384 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-audit-dir\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643411 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3b64526-f2a8-4141-b04c-65b65be09ccc-serving-cert\") pod \"openshift-config-operator-7777fb866f-zhgn2\" (UID: \"c3b64526-f2a8-4141-b04c-65b65be09ccc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zhgn2" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643435 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c3b64526-f2a8-4141-b04c-65b65be09ccc-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zhgn2\" (UID: \"c3b64526-f2a8-4141-b04c-65b65be09ccc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zhgn2" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643456 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mmdhq\" (UID: \"8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmdhq" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643473 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25780238-d742-4461-af80-a4808c4dd40b-config\") pod \"machine-approver-56656f9798-cj64v\" (UID: \"25780238-d742-4461-af80-a4808c4dd40b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cj64v" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643490 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvkg4\" (UniqueName: \"kubernetes.io/projected/25780238-d742-4461-af80-a4808c4dd40b-kube-api-access-xvkg4\") pod \"machine-approver-56656f9798-cj64v\" (UID: \"25780238-d742-4461-af80-a4808c4dd40b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cj64v" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643511 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/66ef6324-33df-4e2a-92dd-36a3f5f7995d-etcd-serving-ca\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643534 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9eab3672-3dd0-4af1-b96d-e8c5e99bb840-stats-auth\") pod \"router-default-5444994796-r7k56\" (UID: \"9eab3672-3dd0-4af1-b96d-e8c5e99bb840\") " pod="openshift-ingress/router-default-5444994796-r7k56" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643554 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643580 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f1b5104-f6eb-4e51-9339-e63f4429f02c-config\") pod \"console-operator-58897d9998-rpd99\" (UID: \"8f1b5104-f6eb-4e51-9339-e63f4429f02c\") " pod="openshift-console-operator/console-operator-58897d9998-rpd99" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643602 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe7c3c56-9f9c-4120-9a89-f840aff6aac8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nl2x7\" (UID: \"fe7c3c56-9f9c-4120-9a89-f840aff6aac8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nl2x7" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643643 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/25780238-d742-4461-af80-a4808c4dd40b-auth-proxy-config\") pod \"machine-approver-56656f9798-cj64v\" (UID: \"25780238-d742-4461-af80-a4808c4dd40b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cj64v" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643663 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643686 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643710 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643735 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjvxr\" (UniqueName: \"kubernetes.io/projected/1bace469-be4b-48d6-9152-d65a4b762f5d-kube-api-access-fjvxr\") pod \"ingress-operator-5b745b69d9-q8lpg\" (UID: \"1bace469-be4b-48d6-9152-d65a4b762f5d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q8lpg" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643756 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-etcd-client\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643780 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe7c3c56-9f9c-4120-9a89-f840aff6aac8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nl2x7\" (UID: \"fe7c3c56-9f9c-4120-9a89-f840aff6aac8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nl2x7" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643802 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/df85fd23-9a76-41a3-b17d-7d06b5bfda87-etcd-ca\") pod \"etcd-operator-b45778765-lxf68\" (UID: \"df85fd23-9a76-41a3-b17d-7d06b5bfda87\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lxf68" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643824 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9eab3672-3dd0-4af1-b96d-e8c5e99bb840-service-ca-bundle\") pod \"router-default-5444994796-r7k56\" (UID: \"9eab3672-3dd0-4af1-b96d-e8c5e99bb840\") " pod="openshift-ingress/router-default-5444994796-r7k56" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643852 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf7f6885-ff5e-44df-ac63-f00fa3ef63a6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-m9m27\" (UID: \"cf7f6885-ff5e-44df-ac63-f00fa3ef63a6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m9m27" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643875 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bg99v\" (UniqueName: \"kubernetes.io/projected/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-kube-api-access-bg99v\") pod \"console-f9d7485db-8r5vw\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643898 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f1a1cfe0-beab-41d4-a478-668f5e7386f9-srv-cert\") pod \"catalog-operator-68c6474976-72wvs\" (UID: \"f1a1cfe0-beab-41d4-a478-668f5e7386f9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-72wvs" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643921 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-console-serving-cert\") pod \"console-f9d7485db-8r5vw\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.643993 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8p5ns\" (UniqueName: \"kubernetes.io/projected/8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4-kube-api-access-8p5ns\") pod \"authentication-operator-69f744f599-mmdhq\" (UID: \"8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmdhq" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.644028 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.644039 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-59cgv"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.644050 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f1b5104-f6eb-4e51-9339-e63f4429f02c-serving-cert\") pod \"console-operator-58897d9998-rpd99\" (UID: \"8f1b5104-f6eb-4e51-9339-e63f4429f02c\") " pod="openshift-console-operator/console-operator-58897d9998-rpd99" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.644138 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.644160 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1bace469-be4b-48d6-9152-d65a4b762f5d-metrics-tls\") pod \"ingress-operator-5b745b69d9-q8lpg\" (UID: \"1bace469-be4b-48d6-9152-d65a4b762f5d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q8lpg" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.644181 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1bace469-be4b-48d6-9152-d65a4b762f5d-trusted-ca\") pod \"ingress-operator-5b745b69d9-q8lpg\" (UID: \"1bace469-be4b-48d6-9152-d65a4b762f5d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q8lpg" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.644205 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf7f6885-ff5e-44df-ac63-f00fa3ef63a6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-m9m27\" (UID: \"cf7f6885-ff5e-44df-ac63-f00fa3ef63a6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m9m27" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.644766 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd0770b5-d9e3-48ae-9ce0-58bcd6a47207-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-rffdv\" (UID: \"dd0770b5-d9e3-48ae-9ce0-58bcd6a47207\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rffdv" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.644812 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-audit-policies\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.644843 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66ef6324-33df-4e2a-92dd-36a3f5f7995d-audit-dir\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.644928 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/df85fd23-9a76-41a3-b17d-7d06b5bfda87-etcd-service-ca\") pod \"etcd-operator-b45778765-lxf68\" (UID: \"df85fd23-9a76-41a3-b17d-7d06b5bfda87\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lxf68" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.645279 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f1b5104-f6eb-4e51-9339-e63f4429f02c-config\") pod \"console-operator-58897d9998-rpd99\" (UID: \"8f1b5104-f6eb-4e51-9339-e63f4429f02c\") " pod="openshift-console-operator/console-operator-58897d9998-rpd99" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.645755 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pvrfw"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.646029 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3ac0673-5554-4ac6-a16e-b08631fdbf94-config\") pod \"controller-manager-879f6c89f-92k7t\" (UID: \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.646409 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.646503 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-config\") pod \"route-controller-manager-6576b87f9c-p86rl\" (UID: \"358682a8-6b18-49ee-bd7d-d9eab2b22ee0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.646935 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/66ef6324-33df-4e2a-92dd-36a3f5f7995d-image-import-ca\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.647345 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/df85fd23-9a76-41a3-b17d-7d06b5bfda87-etcd-ca\") pod \"etcd-operator-b45778765-lxf68\" (UID: \"df85fd23-9a76-41a3-b17d-7d06b5bfda87\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lxf68" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.647773 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe7c3c56-9f9c-4120-9a89-f840aff6aac8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nl2x7\" (UID: \"fe7c3c56-9f9c-4120-9a89-f840aff6aac8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nl2x7" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.648242 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/25780238-d742-4461-af80-a4808c4dd40b-auth-proxy-config\") pod \"machine-approver-56656f9798-cj64v\" (UID: \"25780238-d742-4461-af80-a4808c4dd40b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cj64v" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.649065 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df85fd23-9a76-41a3-b17d-7d06b5bfda87-config\") pod \"etcd-operator-b45778765-lxf68\" (UID: \"df85fd23-9a76-41a3-b17d-7d06b5bfda87\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lxf68" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.649079 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/66ef6324-33df-4e2a-92dd-36a3f5f7995d-encryption-config\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.649188 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66ef6324-33df-4e2a-92dd-36a3f5f7995d-config\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.649719 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1bace469-be4b-48d6-9152-d65a4b762f5d-trusted-ca\") pod \"ingress-operator-5b745b69d9-q8lpg\" (UID: \"1bace469-be4b-48d6-9152-d65a4b762f5d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q8lpg" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.650296 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f3ac0673-5554-4ac6-a16e-b08631fdbf94-client-ca\") pod \"controller-manager-879f6c89f-92k7t\" (UID: \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.650390 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66ef6324-33df-4e2a-92dd-36a3f5f7995d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.650396 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-client-ca\") pod \"route-controller-manager-6576b87f9c-p86rl\" (UID: \"358682a8-6b18-49ee-bd7d-d9eab2b22ee0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.650483 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-audit-dir\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.650713 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/66ef6324-33df-4e2a-92dd-36a3f5f7995d-audit\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.650929 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f1b5104-f6eb-4e51-9339-e63f4429f02c-serving-cert\") pod \"console-operator-58897d9998-rpd99\" (UID: \"8f1b5104-f6eb-4e51-9339-e63f4429f02c\") " pod="openshift-console-operator/console-operator-58897d9998-rpd99" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.651464 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f1b5104-f6eb-4e51-9339-e63f4429f02c-trusted-ca\") pod \"console-operator-58897d9998-rpd99\" (UID: \"8f1b5104-f6eb-4e51-9339-e63f4429f02c\") " pod="openshift-console-operator/console-operator-58897d9998-rpd99" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.651623 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-jx6qd"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.652136 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.652188 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/66ef6324-33df-4e2a-92dd-36a3f5f7995d-node-pullsecrets\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.652408 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/66ef6324-33df-4e2a-92dd-36a3f5f7995d-etcd-client\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.652572 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f3ac0673-5554-4ac6-a16e-b08631fdbf94-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-92k7t\" (UID: \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.652605 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eca6947b-dd5e-449d-a8c5-05277c556bd1-config\") pod \"machine-api-operator-5694c8668f-56j57\" (UID: \"eca6947b-dd5e-449d-a8c5-05277c556bd1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-56j57" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.652713 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/eca6947b-dd5e-449d-a8c5-05277c556bd1-images\") pod \"machine-api-operator-5694c8668f-56j57\" (UID: \"eca6947b-dd5e-449d-a8c5-05277c556bd1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-56j57" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.652886 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5w6mp"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.653447 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.654282 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-svxc5"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.655899 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.655943 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-wlshf"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.656084 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3ac0673-5554-4ac6-a16e-b08631fdbf94-serving-cert\") pod \"controller-manager-879f6c89f-92k7t\" (UID: \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.656732 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/df85fd23-9a76-41a3-b17d-7d06b5bfda87-etcd-client\") pod \"etcd-operator-b45778765-lxf68\" (UID: \"df85fd23-9a76-41a3-b17d-7d06b5bfda87\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lxf68" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.656842 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.656981 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-gzp8q"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.657432 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.658079 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1bace469-be4b-48d6-9152-d65a4b762f5d-metrics-tls\") pod \"ingress-operator-5b745b69d9-q8lpg\" (UID: \"1bace469-be4b-48d6-9152-d65a4b762f5d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q8lpg" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.658369 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399445-bkt2p"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.658389 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.658744 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/25780238-d742-4461-af80-a4808c4dd40b-machine-approver-tls\") pod \"machine-approver-56656f9798-cj64v\" (UID: \"25780238-d742-4461-af80-a4808c4dd40b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cj64v" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.658789 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd0770b5-d9e3-48ae-9ce0-58bcd6a47207-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-rffdv\" (UID: \"dd0770b5-d9e3-48ae-9ce0-58bcd6a47207\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rffdv" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.658831 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-audit-dir\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.659080 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df85fd23-9a76-41a3-b17d-7d06b5bfda87-serving-cert\") pod \"etcd-operator-b45778765-lxf68\" (UID: \"df85fd23-9a76-41a3-b17d-7d06b5bfda87\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lxf68" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.659301 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-audit-policies\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.659352 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.660396 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25780238-d742-4461-af80-a4808c4dd40b-config\") pod \"machine-approver-56656f9798-cj64v\" (UID: \"25780238-d742-4461-af80-a4808c4dd40b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cj64v" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.660403 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.660660 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-cb66b"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.661112 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-encryption-config\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.661176 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-serving-cert\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.661213 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-cb66b" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.661454 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66ef6324-33df-4e2a-92dd-36a3f5f7995d-serving-cert\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.661504 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.661805 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/66ef6324-33df-4e2a-92dd-36a3f5f7995d-etcd-serving-ca\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.661828 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.661845 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-cb66b"] Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.661899 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/eca6947b-dd5e-449d-a8c5-05277c556bd1-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-56j57\" (UID: \"eca6947b-dd5e-449d-a8c5-05277c556bd1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-56j57" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.662533 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-etcd-client\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.662728 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-serving-cert\") pod \"route-controller-manager-6576b87f9c-p86rl\" (UID: \"358682a8-6b18-49ee-bd7d-d9eab2b22ee0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.662753 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.664895 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe7c3c56-9f9c-4120-9a89-f840aff6aac8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nl2x7\" (UID: \"fe7c3c56-9f9c-4120-9a89-f840aff6aac8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nl2x7" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.670152 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.689496 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.709793 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.729770 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.744878 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-service-ca\") pod \"console-f9d7485db-8r5vw\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.744908 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4-serving-cert\") pod \"authentication-operator-69f744f599-mmdhq\" (UID: \"8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmdhq" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.744929 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r755l\" (UniqueName: \"kubernetes.io/projected/f1a1cfe0-beab-41d4-a478-668f5e7386f9-kube-api-access-r755l\") pod \"catalog-operator-68c6474976-72wvs\" (UID: \"f1a1cfe0-beab-41d4-a478-668f5e7386f9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-72wvs" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.744953 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmffc\" (UniqueName: \"kubernetes.io/projected/cf7f6885-ff5e-44df-ac63-f00fa3ef63a6-kube-api-access-qmffc\") pod \"openshift-apiserver-operator-796bbdcf4f-m9m27\" (UID: \"cf7f6885-ff5e-44df-ac63-f00fa3ef63a6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m9m27" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.744991 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/7bf55845-293a-497e-87f1-eb5c36670105-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-mqs58\" (UID: \"7bf55845-293a-497e-87f1-eb5c36670105\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mqs58" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745010 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfjgk\" (UniqueName: \"kubernetes.io/projected/4e3c9b7d-5fa3-4f70-babe-522cb8dc7443-kube-api-access-rfjgk\") pod \"cluster-image-registry-operator-dc59b4c8b-q6jp2\" (UID: \"4e3c9b7d-5fa3-4f70-babe-522cb8dc7443\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6jp2" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745029 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/4e3c9b7d-5fa3-4f70-babe-522cb8dc7443-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-q6jp2\" (UID: \"4e3c9b7d-5fa3-4f70-babe-522cb8dc7443\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6jp2" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745061 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f1a1cfe0-beab-41d4-a478-668f5e7386f9-profile-collector-cert\") pod \"catalog-operator-68c6474976-72wvs\" (UID: \"f1a1cfe0-beab-41d4-a478-668f5e7386f9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-72wvs" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745082 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq84x\" (UniqueName: \"kubernetes.io/projected/7bf55845-293a-497e-87f1-eb5c36670105-kube-api-access-hq84x\") pod \"cluster-samples-operator-665b6dd947-mqs58\" (UID: \"7bf55845-293a-497e-87f1-eb5c36670105\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mqs58" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745106 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-console-config\") pod \"console-f9d7485db-8r5vw\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745130 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntbxx\" (UniqueName: \"kubernetes.io/projected/3e14d56e-43b3-40ec-809f-371e8ab2f628-kube-api-access-ntbxx\") pod \"control-plane-machine-set-operator-78cbb6b69f-xfdvj\" (UID: \"3e14d56e-43b3-40ec-809f-371e8ab2f628\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xfdvj" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745155 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-console-oauth-config\") pod \"console-f9d7485db-8r5vw\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745185 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-trusted-ca-bundle\") pod \"console-f9d7485db-8r5vw\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745206 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4e3c9b7d-5fa3-4f70-babe-522cb8dc7443-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-q6jp2\" (UID: \"4e3c9b7d-5fa3-4f70-babe-522cb8dc7443\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6jp2" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745229 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4-service-ca-bundle\") pod \"authentication-operator-69f744f599-mmdhq\" (UID: \"8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmdhq" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745254 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4e3c9b7d-5fa3-4f70-babe-522cb8dc7443-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-q6jp2\" (UID: \"4e3c9b7d-5fa3-4f70-babe-522cb8dc7443\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6jp2" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745276 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8b246b95-be2e-4c90-8cf0-f92c612bfbc1-srv-cert\") pod \"olm-operator-6b444d44fb-pmsgp\" (UID: \"8b246b95-be2e-4c90-8cf0-f92c612bfbc1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pmsgp" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745306 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9eab3672-3dd0-4af1-b96d-e8c5e99bb840-default-certificate\") pod \"router-default-5444994796-r7k56\" (UID: \"9eab3672-3dd0-4af1-b96d-e8c5e99bb840\") " pod="openshift-ingress/router-default-5444994796-r7k56" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745328 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-oauth-serving-cert\") pod \"console-f9d7485db-8r5vw\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745365 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9eab3672-3dd0-4af1-b96d-e8c5e99bb840-metrics-certs\") pod \"router-default-5444994796-r7k56\" (UID: \"9eab3672-3dd0-4af1-b96d-e8c5e99bb840\") " pod="openshift-ingress/router-default-5444994796-r7k56" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745407 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mmdhq\" (UID: \"8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmdhq" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745431 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3b64526-f2a8-4141-b04c-65b65be09ccc-serving-cert\") pod \"openshift-config-operator-7777fb866f-zhgn2\" (UID: \"c3b64526-f2a8-4141-b04c-65b65be09ccc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zhgn2" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745454 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c3b64526-f2a8-4141-b04c-65b65be09ccc-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zhgn2\" (UID: \"c3b64526-f2a8-4141-b04c-65b65be09ccc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zhgn2" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745484 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9eab3672-3dd0-4af1-b96d-e8c5e99bb840-stats-auth\") pod \"router-default-5444994796-r7k56\" (UID: \"9eab3672-3dd0-4af1-b96d-e8c5e99bb840\") " pod="openshift-ingress/router-default-5444994796-r7k56" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745516 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9eab3672-3dd0-4af1-b96d-e8c5e99bb840-service-ca-bundle\") pod \"router-default-5444994796-r7k56\" (UID: \"9eab3672-3dd0-4af1-b96d-e8c5e99bb840\") " pod="openshift-ingress/router-default-5444994796-r7k56" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745539 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf7f6885-ff5e-44df-ac63-f00fa3ef63a6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-m9m27\" (UID: \"cf7f6885-ff5e-44df-ac63-f00fa3ef63a6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m9m27" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745561 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bg99v\" (UniqueName: \"kubernetes.io/projected/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-kube-api-access-bg99v\") pod \"console-f9d7485db-8r5vw\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745585 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f1a1cfe0-beab-41d4-a478-668f5e7386f9-srv-cert\") pod \"catalog-operator-68c6474976-72wvs\" (UID: \"f1a1cfe0-beab-41d4-a478-668f5e7386f9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-72wvs" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745606 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-console-serving-cert\") pod \"console-f9d7485db-8r5vw\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745628 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8p5ns\" (UniqueName: \"kubernetes.io/projected/8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4-kube-api-access-8p5ns\") pod \"authentication-operator-69f744f599-mmdhq\" (UID: \"8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmdhq" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745652 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf7f6885-ff5e-44df-ac63-f00fa3ef63a6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-m9m27\" (UID: \"cf7f6885-ff5e-44df-ac63-f00fa3ef63a6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m9m27" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745689 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdr48\" (UniqueName: \"kubernetes.io/projected/9eab3672-3dd0-4af1-b96d-e8c5e99bb840-kube-api-access-hdr48\") pod \"router-default-5444994796-r7k56\" (UID: \"9eab3672-3dd0-4af1-b96d-e8c5e99bb840\") " pod="openshift-ingress/router-default-5444994796-r7k56" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745713 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmvs7\" (UniqueName: \"kubernetes.io/projected/8b246b95-be2e-4c90-8cf0-f92c612bfbc1-kube-api-access-lmvs7\") pod \"olm-operator-6b444d44fb-pmsgp\" (UID: \"8b246b95-be2e-4c90-8cf0-f92c612bfbc1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pmsgp" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745735 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4-config\") pod \"authentication-operator-69f744f599-mmdhq\" (UID: \"8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmdhq" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745778 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8b246b95-be2e-4c90-8cf0-f92c612bfbc1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-pmsgp\" (UID: \"8b246b95-be2e-4c90-8cf0-f92c612bfbc1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pmsgp" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745802 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtrcx\" (UniqueName: \"kubernetes.io/projected/c3b64526-f2a8-4141-b04c-65b65be09ccc-kube-api-access-xtrcx\") pod \"openshift-config-operator-7777fb866f-zhgn2\" (UID: \"c3b64526-f2a8-4141-b04c-65b65be09ccc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zhgn2" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745854 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3e14d56e-43b3-40ec-809f-371e8ab2f628-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-xfdvj\" (UID: \"3e14d56e-43b3-40ec-809f-371e8ab2f628\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xfdvj" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.745899 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fc65\" (UniqueName: \"kubernetes.io/projected/96284acd-e6e2-467c-8c70-005ca153dda3-kube-api-access-2fc65\") pod \"migrator-59844c95c7-54rmc\" (UID: \"96284acd-e6e2-467c-8c70-005ca153dda3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-54rmc" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.746447 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-console-config\") pod \"console-f9d7485db-8r5vw\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.746720 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c3b64526-f2a8-4141-b04c-65b65be09ccc-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zhgn2\" (UID: \"c3b64526-f2a8-4141-b04c-65b65be09ccc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zhgn2" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.746759 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4-config\") pod \"authentication-operator-69f744f599-mmdhq\" (UID: \"8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmdhq" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.746956 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9eab3672-3dd0-4af1-b96d-e8c5e99bb840-service-ca-bundle\") pod \"router-default-5444994796-r7k56\" (UID: \"9eab3672-3dd0-4af1-b96d-e8c5e99bb840\") " pod="openshift-ingress/router-default-5444994796-r7k56" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.747413 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-service-ca\") pod \"console-f9d7485db-8r5vw\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.747780 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4-serving-cert\") pod \"authentication-operator-69f744f599-mmdhq\" (UID: \"8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmdhq" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.748120 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-trusted-ca-bundle\") pod \"console-f9d7485db-8r5vw\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.748166 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-oauth-serving-cert\") pod \"console-f9d7485db-8r5vw\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.749496 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9eab3672-3dd0-4af1-b96d-e8c5e99bb840-default-certificate\") pod \"router-default-5444994796-r7k56\" (UID: \"9eab3672-3dd0-4af1-b96d-e8c5e99bb840\") " pod="openshift-ingress/router-default-5444994796-r7k56" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.749522 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9eab3672-3dd0-4af1-b96d-e8c5e99bb840-stats-auth\") pod \"router-default-5444994796-r7k56\" (UID: \"9eab3672-3dd0-4af1-b96d-e8c5e99bb840\") " pod="openshift-ingress/router-default-5444994796-r7k56" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.749936 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-console-oauth-config\") pod \"console-f9d7485db-8r5vw\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.750025 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-console-serving-cert\") pod \"console-f9d7485db-8r5vw\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.752850 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9eab3672-3dd0-4af1-b96d-e8c5e99bb840-metrics-certs\") pod \"router-default-5444994796-r7k56\" (UID: \"9eab3672-3dd0-4af1-b96d-e8c5e99bb840\") " pod="openshift-ingress/router-default-5444994796-r7k56" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.757033 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.767707 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mmdhq\" (UID: \"8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmdhq" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.769747 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.777209 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4-service-ca-bundle\") pod \"authentication-operator-69f744f599-mmdhq\" (UID: \"8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmdhq" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.790711 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.809793 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.830603 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.853690 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.870061 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.890476 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.914160 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.919047 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf7f6885-ff5e-44df-ac63-f00fa3ef63a6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-m9m27\" (UID: \"cf7f6885-ff5e-44df-ac63-f00fa3ef63a6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m9m27" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.929808 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.936930 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf7f6885-ff5e-44df-ac63-f00fa3ef63a6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-m9m27\" (UID: \"cf7f6885-ff5e-44df-ac63-f00fa3ef63a6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m9m27" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.949354 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.970551 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.980192 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3b64526-f2a8-4141-b04c-65b65be09ccc-serving-cert\") pod \"openshift-config-operator-7777fb866f-zhgn2\" (UID: \"c3b64526-f2a8-4141-b04c-65b65be09ccc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zhgn2" Nov 24 06:57:19 crc kubenswrapper[4809]: I1124 06:57:19.989982 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.010511 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.030409 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.050284 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.069728 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.089742 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.110256 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.131318 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.150315 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.162125 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3e14d56e-43b3-40ec-809f-371e8ab2f628-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-xfdvj\" (UID: \"3e14d56e-43b3-40ec-809f-371e8ab2f628\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xfdvj" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.169633 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.189777 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.219511 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.228158 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4e3c9b7d-5fa3-4f70-babe-522cb8dc7443-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-q6jp2\" (UID: \"4e3c9b7d-5fa3-4f70-babe-522cb8dc7443\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6jp2" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.230038 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.250866 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.270092 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.290756 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.309819 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.318456 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/4e3c9b7d-5fa3-4f70-babe-522cb8dc7443-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-q6jp2\" (UID: \"4e3c9b7d-5fa3-4f70-babe-522cb8dc7443\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6jp2" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.330011 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.350040 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.370173 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.390141 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.410007 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.431468 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.440191 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f1a1cfe0-beab-41d4-a478-668f5e7386f9-srv-cert\") pod \"catalog-operator-68c6474976-72wvs\" (UID: \"f1a1cfe0-beab-41d4-a478-668f5e7386f9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-72wvs" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.451553 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.470771 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.479510 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f1a1cfe0-beab-41d4-a478-668f5e7386f9-profile-collector-cert\") pod \"catalog-operator-68c6474976-72wvs\" (UID: \"f1a1cfe0-beab-41d4-a478-668f5e7386f9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-72wvs" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.480232 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8b246b95-be2e-4c90-8cf0-f92c612bfbc1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-pmsgp\" (UID: \"8b246b95-be2e-4c90-8cf0-f92c612bfbc1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pmsgp" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.490233 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.510523 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.529921 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.550367 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.570626 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.580533 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8b246b95-be2e-4c90-8cf0-f92c612bfbc1-srv-cert\") pod \"olm-operator-6b444d44fb-pmsgp\" (UID: \"8b246b95-be2e-4c90-8cf0-f92c612bfbc1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pmsgp" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.589274 4809 request.go:700] Waited for 1.014939035s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/secrets?fieldSelector=metadata.name%3Dimage-registry-tls&limit=500&resourceVersion=0 Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.590957 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.612137 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.630487 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.650477 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.670281 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.691028 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.701459 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/7bf55845-293a-497e-87f1-eb5c36670105-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-mqs58\" (UID: \"7bf55845-293a-497e-87f1-eb5c36670105\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mqs58" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.709884 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.750631 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.771261 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.790220 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.810312 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.830359 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.850263 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.870799 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.890852 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.909923 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.930999 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.951165 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.970333 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 24 06:57:20 crc kubenswrapper[4809]: I1124 06:57:20.990147 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.010257 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.030209 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.051038 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.070827 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.091043 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.112327 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.130513 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.150076 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.171312 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.199078 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.212437 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.230849 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.250317 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.270328 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.289781 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.312004 4809 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.330135 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.350610 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.371318 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.408149 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr6gp\" (UniqueName: \"kubernetes.io/projected/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-kube-api-access-wr6gp\") pod \"route-controller-manager-6576b87f9c-p86rl\" (UID: \"358682a8-6b18-49ee-bd7d-d9eab2b22ee0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.431790 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8frv\" (UniqueName: \"kubernetes.io/projected/8f1b5104-f6eb-4e51-9339-e63f4429f02c-kube-api-access-z8frv\") pod \"console-operator-58897d9998-rpd99\" (UID: \"8f1b5104-f6eb-4e51-9339-e63f4429f02c\") " pod="openshift-console-operator/console-operator-58897d9998-rpd99" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.448349 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r4sq\" (UniqueName: \"kubernetes.io/projected/f3ac0673-5554-4ac6-a16e-b08631fdbf94-kube-api-access-4r4sq\") pod \"controller-manager-879f6c89f-92k7t\" (UID: \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.451191 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-rpd99" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.470386 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1bace469-be4b-48d6-9152-d65a4b762f5d-bound-sa-token\") pod \"ingress-operator-5b745b69d9-q8lpg\" (UID: \"1bace469-be4b-48d6-9152-d65a4b762f5d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q8lpg" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.494529 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4mvw\" (UniqueName: \"kubernetes.io/projected/df85fd23-9a76-41a3-b17d-7d06b5bfda87-kube-api-access-m4mvw\") pod \"etcd-operator-b45778765-lxf68\" (UID: \"df85fd23-9a76-41a3-b17d-7d06b5bfda87\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lxf68" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.503684 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbmcq\" (UniqueName: \"kubernetes.io/projected/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-kube-api-access-rbmcq\") pod \"oauth-openshift-558db77b4-69dgx\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.524081 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nh4b\" (UniqueName: \"kubernetes.io/projected/fe7c3c56-9f9c-4120-9a89-f840aff6aac8-kube-api-access-8nh4b\") pod \"kube-storage-version-migrator-operator-b67b599dd-nl2x7\" (UID: \"fe7c3c56-9f9c-4120-9a89-f840aff6aac8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nl2x7" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.556862 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7vgf\" (UniqueName: \"kubernetes.io/projected/eca6947b-dd5e-449d-a8c5-05277c556bd1-kube-api-access-l7vgf\") pod \"machine-api-operator-5694c8668f-56j57\" (UID: \"eca6947b-dd5e-449d-a8c5-05277c556bd1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-56j57" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.576102 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmwcj\" (UniqueName: \"kubernetes.io/projected/b6237cc7-ba93-45a6-b560-d7916abf71f7-kube-api-access-mmwcj\") pod \"downloads-7954f5f757-lkdj2\" (UID: \"b6237cc7-ba93-45a6-b560-d7916abf71f7\") " pod="openshift-console/downloads-7954f5f757-lkdj2" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.596672 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjvxr\" (UniqueName: \"kubernetes.io/projected/1bace469-be4b-48d6-9152-d65a4b762f5d-kube-api-access-fjvxr\") pod \"ingress-operator-5b745b69d9-q8lpg\" (UID: \"1bace469-be4b-48d6-9152-d65a4b762f5d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q8lpg" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.600820 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.604694 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhczn\" (UniqueName: \"kubernetes.io/projected/66ef6324-33df-4e2a-92dd-36a3f5f7995d-kube-api-access-dhczn\") pod \"apiserver-76f77b778f-88wkc\" (UID: \"66ef6324-33df-4e2a-92dd-36a3f5f7995d\") " pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.609080 4809 request.go:700] Waited for 1.958574974s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-oauth-apiserver/serviceaccounts/oauth-apiserver-sa/token Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.628086 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxqtp\" (UniqueName: \"kubernetes.io/projected/04b0f650-8f86-4d8e-9fbc-860516a0f3a7-kube-api-access-sxqtp\") pod \"apiserver-7bbb656c7d-88j9l\" (UID: \"04b0f650-8f86-4d8e-9fbc-860516a0f3a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.634140 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rpd99"] Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.642072 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.644380 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj2x5\" (UniqueName: \"kubernetes.io/projected/dd0770b5-d9e3-48ae-9ce0-58bcd6a47207-kube-api-access-dj2x5\") pod \"openshift-controller-manager-operator-756b6f6bc6-rffdv\" (UID: \"dd0770b5-d9e3-48ae-9ce0-58bcd6a47207\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rffdv" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.647264 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.673204 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvkg4\" (UniqueName: \"kubernetes.io/projected/25780238-d742-4461-af80-a4808c4dd40b-kube-api-access-xvkg4\") pod \"machine-approver-56656f9798-cj64v\" (UID: \"25780238-d742-4461-af80-a4808c4dd40b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cj64v" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.678226 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-lkdj2" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.688841 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.691365 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.698354 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.703424 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rffdv" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.711169 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.717517 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-lxf68" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.724548 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nl2x7" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.736407 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.739443 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q8lpg" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.753294 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.805905 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfjgk\" (UniqueName: \"kubernetes.io/projected/4e3c9b7d-5fa3-4f70-babe-522cb8dc7443-kube-api-access-rfjgk\") pod \"cluster-image-registry-operator-dc59b4c8b-q6jp2\" (UID: \"4e3c9b7d-5fa3-4f70-babe-522cb8dc7443\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6jp2" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.807265 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmffc\" (UniqueName: \"kubernetes.io/projected/cf7f6885-ff5e-44df-ac63-f00fa3ef63a6-kube-api-access-qmffc\") pod \"openshift-apiserver-operator-796bbdcf4f-m9m27\" (UID: \"cf7f6885-ff5e-44df-ac63-f00fa3ef63a6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m9m27" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.833599 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-92k7t"] Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.837286 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r755l\" (UniqueName: \"kubernetes.io/projected/f1a1cfe0-beab-41d4-a478-668f5e7386f9-kube-api-access-r755l\") pod \"catalog-operator-68c6474976-72wvs\" (UID: \"f1a1cfe0-beab-41d4-a478-668f5e7386f9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-72wvs" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.844189 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m9m27" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.849455 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4e3c9b7d-5fa3-4f70-babe-522cb8dc7443-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-q6jp2\" (UID: \"4e3c9b7d-5fa3-4f70-babe-522cb8dc7443\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6jp2" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.855199 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-56j57" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.887951 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntbxx\" (UniqueName: \"kubernetes.io/projected/3e14d56e-43b3-40ec-809f-371e8ab2f628-kube-api-access-ntbxx\") pod \"control-plane-machine-set-operator-78cbb6b69f-xfdvj\" (UID: \"3e14d56e-43b3-40ec-809f-371e8ab2f628\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xfdvj" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.896072 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6jp2" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.897500 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq84x\" (UniqueName: \"kubernetes.io/projected/7bf55845-293a-497e-87f1-eb5c36670105-kube-api-access-hq84x\") pod \"cluster-samples-operator-665b6dd947-mqs58\" (UID: \"7bf55845-293a-497e-87f1-eb5c36670105\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mqs58" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.904406 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l"] Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.904565 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdr48\" (UniqueName: \"kubernetes.io/projected/9eab3672-3dd0-4af1-b96d-e8c5e99bb840-kube-api-access-hdr48\") pod \"router-default-5444994796-r7k56\" (UID: \"9eab3672-3dd0-4af1-b96d-e8c5e99bb840\") " pod="openshift-ingress/router-default-5444994796-r7k56" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.912193 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-72wvs" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.927164 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fc65\" (UniqueName: \"kubernetes.io/projected/96284acd-e6e2-467c-8c70-005ca153dda3-kube-api-access-2fc65\") pod \"migrator-59844c95c7-54rmc\" (UID: \"96284acd-e6e2-467c-8c70-005ca153dda3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-54rmc" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.944495 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bg99v\" (UniqueName: \"kubernetes.io/projected/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-kube-api-access-bg99v\") pod \"console-f9d7485db-8r5vw\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.958289 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mqs58" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.965880 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmvs7\" (UniqueName: \"kubernetes.io/projected/8b246b95-be2e-4c90-8cf0-f92c612bfbc1-kube-api-access-lmvs7\") pod \"olm-operator-6b444d44fb-pmsgp\" (UID: \"8b246b95-be2e-4c90-8cf0-f92c612bfbc1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pmsgp" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.966013 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cj64v" Nov 24 06:57:21 crc kubenswrapper[4809]: I1124 06:57:21.987365 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtrcx\" (UniqueName: \"kubernetes.io/projected/c3b64526-f2a8-4141-b04c-65b65be09ccc-kube-api-access-xtrcx\") pod \"openshift-config-operator-7777fb866f-zhgn2\" (UID: \"c3b64526-f2a8-4141-b04c-65b65be09ccc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zhgn2" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.013079 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8p5ns\" (UniqueName: \"kubernetes.io/projected/8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4-kube-api-access-8p5ns\") pod \"authentication-operator-69f744f599-mmdhq\" (UID: \"8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmdhq" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.028018 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-lxf68"] Nov 24 06:57:22 crc kubenswrapper[4809]: W1124 06:57:22.054907 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf85fd23_9a76_41a3_b17d_7d06b5bfda87.slice/crio-aa28271282722b226d613866756086abbd31320beba7b3e99c8e0c73fc94e537 WatchSource:0}: Error finding container aa28271282722b226d613866756086abbd31320beba7b3e99c8e0c73fc94e537: Status 404 returned error can't find the container with id aa28271282722b226d613866756086abbd31320beba7b3e99c8e0c73fc94e537 Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.084677 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4ce90879-8e80-46ec-9f0b-f972764812da-metrics-tls\") pod \"dns-operator-744455d44c-q9xpc\" (UID: \"4ce90879-8e80-46ec-9f0b-f972764812da\") " pod="openshift-dns-operator/dns-operator-744455d44c-q9xpc" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.084716 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4af1418f-a56e-4f26-a800-3b402b7da053-installation-pull-secrets\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.084744 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrpmg\" (UniqueName: \"kubernetes.io/projected/4ce90879-8e80-46ec-9f0b-f972764812da-kube-api-access-wrpmg\") pod \"dns-operator-744455d44c-q9xpc\" (UID: \"4ce90879-8e80-46ec-9f0b-f972764812da\") " pod="openshift-dns-operator/dns-operator-744455d44c-q9xpc" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.084760 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4af1418f-a56e-4f26-a800-3b402b7da053-ca-trust-extracted\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.084780 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c002f86-a984-4feb-b110-35aebfee516e-config\") pod \"kube-controller-manager-operator-78b949d7b-6ggzn\" (UID: \"1c002f86-a984-4feb-b110-35aebfee516e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ggzn" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.084805 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd9f4c00-90bd-4782-8ceb-eaab6a9799c3-config\") pod \"kube-apiserver-operator-766d6c64bb-prdvb\" (UID: \"cd9f4c00-90bd-4782-8ceb-eaab6a9799c3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prdvb" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.084819 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cd9f4c00-90bd-4782-8ceb-eaab6a9799c3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-prdvb\" (UID: \"cd9f4c00-90bd-4782-8ceb-eaab6a9799c3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prdvb" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.084851 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.084877 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnc8h\" (UniqueName: \"kubernetes.io/projected/c07f2b54-9258-4f6d-bf9f-7f28e4331a84-kube-api-access-gnc8h\") pod \"machine-config-operator-74547568cd-7s9hl\" (UID: \"c07f2b54-9258-4f6d-bf9f-7f28e4331a84\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7s9hl" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.084897 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gl9h\" (UniqueName: \"kubernetes.io/projected/4af1418f-a56e-4f26-a800-3b402b7da053-kube-api-access-9gl9h\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.084936 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4af1418f-a56e-4f26-a800-3b402b7da053-trusted-ca\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.084959 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4af1418f-a56e-4f26-a800-3b402b7da053-registry-tls\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.085010 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c002f86-a984-4feb-b110-35aebfee516e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6ggzn\" (UID: \"1c002f86-a984-4feb-b110-35aebfee516e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ggzn" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.085075 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4af1418f-a56e-4f26-a800-3b402b7da053-bound-sa-token\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.085093 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1c002f86-a984-4feb-b110-35aebfee516e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6ggzn\" (UID: \"1c002f86-a984-4feb-b110-35aebfee516e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ggzn" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.085131 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e3ba6e1a-6434-4867-8a19-9c9a1008b688-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xq929\" (UID: \"e3ba6e1a-6434-4867-8a19-9c9a1008b688\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xq929" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.085173 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3ba6e1a-6434-4867-8a19-9c9a1008b688-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xq929\" (UID: \"e3ba6e1a-6434-4867-8a19-9c9a1008b688\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xq929" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.085187 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd9f4c00-90bd-4782-8ceb-eaab6a9799c3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-prdvb\" (UID: \"cd9f4c00-90bd-4782-8ceb-eaab6a9799c3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prdvb" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.085201 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c07f2b54-9258-4f6d-bf9f-7f28e4331a84-images\") pod \"machine-config-operator-74547568cd-7s9hl\" (UID: \"c07f2b54-9258-4f6d-bf9f-7f28e4331a84\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7s9hl" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.085229 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4af1418f-a56e-4f26-a800-3b402b7da053-registry-certificates\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.085257 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c07f2b54-9258-4f6d-bf9f-7f28e4331a84-proxy-tls\") pod \"machine-config-operator-74547568cd-7s9hl\" (UID: \"c07f2b54-9258-4f6d-bf9f-7f28e4331a84\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7s9hl" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.085272 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3ba6e1a-6434-4867-8a19-9c9a1008b688-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xq929\" (UID: \"e3ba6e1a-6434-4867-8a19-9c9a1008b688\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xq929" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.085300 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c07f2b54-9258-4f6d-bf9f-7f28e4331a84-auth-proxy-config\") pod \"machine-config-operator-74547568cd-7s9hl\" (UID: \"c07f2b54-9258-4f6d-bf9f-7f28e4331a84\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7s9hl" Nov 24 06:57:22 crc kubenswrapper[4809]: E1124 06:57:22.086062 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:22.586044349 +0000 UTC m=+142.486636164 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.106946 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-r7k56" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.114130 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.121217 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-54rmc" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.129287 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-mmdhq" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.130738 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-lkdj2"] Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.133265 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zhgn2" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.134457 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-88wkc"] Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.166007 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-q8lpg"] Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.172921 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rffdv"] Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.180595 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xfdvj" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.186580 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:22 crc kubenswrapper[4809]: E1124 06:57:22.186725 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:22.686704443 +0000 UTC m=+142.587296148 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.186815 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/47242c3c-e4f8-49f4-b508-e705396600b2-socket-dir\") pod \"csi-hostpathplugin-gzp8q\" (UID: \"47242c3c-e4f8-49f4-b508-e705396600b2\") " pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.186845 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/50767fcc-443e-40ce-84df-2182a03691f1-tmpfs\") pod \"packageserver-d55dfcdfc-r5mlr\" (UID: \"50767fcc-443e-40ce-84df-2182a03691f1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.186893 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4ce90879-8e80-46ec-9f0b-f972764812da-metrics-tls\") pod \"dns-operator-744455d44c-q9xpc\" (UID: \"4ce90879-8e80-46ec-9f0b-f972764812da\") " pod="openshift-dns-operator/dns-operator-744455d44c-q9xpc" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.186931 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b08de060-f5b7-4bed-a27c-89294a8827fa-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-svxc5\" (UID: \"b08de060-f5b7-4bed-a27c-89294a8827fa\") " pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.186953 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c1a058e6-56e0-4020-a9aa-e6537e1ccc39-signing-key\") pod \"service-ca-9c57cc56f-z2dgw\" (UID: \"c1a058e6-56e0-4020-a9aa-e6537e1ccc39\") " pod="openshift-service-ca/service-ca-9c57cc56f-z2dgw" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187004 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d0b8935a-d42d-4f07-876b-67f7e1ce3d89-proxy-tls\") pod \"machine-config-controller-84d6567774-jzwg5\" (UID: \"d0b8935a-d42d-4f07-876b-67f7e1ce3d89\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzwg5" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187034 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e846497c-2339-4f20-b688-f2483c07d54f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pvrfw\" (UID: \"e846497c-2339-4f20-b688-f2483c07d54f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pvrfw" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187087 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/50767fcc-443e-40ce-84df-2182a03691f1-apiservice-cert\") pod \"packageserver-d55dfcdfc-r5mlr\" (UID: \"50767fcc-443e-40ce-84df-2182a03691f1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187114 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4af1418f-a56e-4f26-a800-3b402b7da053-installation-pull-secrets\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187130 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/476eceb4-f8fa-49c1-9015-5e067998af80-secret-volume\") pod \"collect-profiles-29399445-bkt2p\" (UID: \"476eceb4-f8fa-49c1-9015-5e067998af80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-bkt2p" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187146 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj72f\" (UniqueName: \"kubernetes.io/projected/053a3f86-3670-42c2-afe9-8b1f09ece542-kube-api-access-jj72f\") pod \"dns-default-wlshf\" (UID: \"053a3f86-3670-42c2-afe9-8b1f09ece542\") " pod="openshift-dns/dns-default-wlshf" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187161 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpgps\" (UniqueName: \"kubernetes.io/projected/e846497c-2339-4f20-b688-f2483c07d54f-kube-api-access-lpgps\") pod \"multus-admission-controller-857f4d67dd-pvrfw\" (UID: \"e846497c-2339-4f20-b688-f2483c07d54f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pvrfw" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187176 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dm64\" (UniqueName: \"kubernetes.io/projected/c1a058e6-56e0-4020-a9aa-e6537e1ccc39-kube-api-access-4dm64\") pod \"service-ca-9c57cc56f-z2dgw\" (UID: \"c1a058e6-56e0-4020-a9aa-e6537e1ccc39\") " pod="openshift-service-ca/service-ca-9c57cc56f-z2dgw" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187209 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4af1418f-a56e-4f26-a800-3b402b7da053-ca-trust-extracted\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187229 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrpmg\" (UniqueName: \"kubernetes.io/projected/4ce90879-8e80-46ec-9f0b-f972764812da-kube-api-access-wrpmg\") pod \"dns-operator-744455d44c-q9xpc\" (UID: \"4ce90879-8e80-46ec-9f0b-f972764812da\") " pod="openshift-dns-operator/dns-operator-744455d44c-q9xpc" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187246 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhrdn\" (UniqueName: \"kubernetes.io/projected/34fdb8d0-057e-40b9-a765-3e6a456ed218-kube-api-access-vhrdn\") pod \"ingress-canary-cb66b\" (UID: \"34fdb8d0-057e-40b9-a765-3e6a456ed218\") " pod="openshift-ingress-canary/ingress-canary-cb66b" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187333 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c002f86-a984-4feb-b110-35aebfee516e-config\") pod \"kube-controller-manager-operator-78b949d7b-6ggzn\" (UID: \"1c002f86-a984-4feb-b110-35aebfee516e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ggzn" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187400 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/50767fcc-443e-40ce-84df-2182a03691f1-webhook-cert\") pod \"packageserver-d55dfcdfc-r5mlr\" (UID: \"50767fcc-443e-40ce-84df-2182a03691f1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187458 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jf5tz\" (UniqueName: \"kubernetes.io/projected/e302b261-06bb-4ed0-9479-c326f26bfe43-kube-api-access-jf5tz\") pod \"package-server-manager-789f6589d5-5w6mp\" (UID: \"e302b261-06bb-4ed0-9479-c326f26bfe43\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5w6mp" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187513 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd9f4c00-90bd-4782-8ceb-eaab6a9799c3-config\") pod \"kube-apiserver-operator-766d6c64bb-prdvb\" (UID: \"cd9f4c00-90bd-4782-8ceb-eaab6a9799c3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prdvb" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187531 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cd9f4c00-90bd-4782-8ceb-eaab6a9799c3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-prdvb\" (UID: \"cd9f4c00-90bd-4782-8ceb-eaab6a9799c3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prdvb" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187549 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/47242c3c-e4f8-49f4-b508-e705396600b2-plugins-dir\") pod \"csi-hostpathplugin-gzp8q\" (UID: \"47242c3c-e4f8-49f4-b508-e705396600b2\") " pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187586 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szl84\" (UniqueName: \"kubernetes.io/projected/476eceb4-f8fa-49c1-9015-5e067998af80-kube-api-access-szl84\") pod \"collect-profiles-29399445-bkt2p\" (UID: \"476eceb4-f8fa-49c1-9015-5e067998af80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-bkt2p" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187687 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187778 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgqft\" (UniqueName: \"kubernetes.io/projected/3cf1e8ba-0d51-4ccf-bdf1-24959b8c5c31-kube-api-access-xgqft\") pod \"machine-config-server-tk28k\" (UID: \"3cf1e8ba-0d51-4ccf-bdf1-24959b8c5c31\") " pod="openshift-machine-config-operator/machine-config-server-tk28k" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187867 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnc8h\" (UniqueName: \"kubernetes.io/projected/c07f2b54-9258-4f6d-bf9f-7f28e4331a84-kube-api-access-gnc8h\") pod \"machine-config-operator-74547568cd-7s9hl\" (UID: \"c07f2b54-9258-4f6d-bf9f-7f28e4331a84\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7s9hl" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187906 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/47242c3c-e4f8-49f4-b508-e705396600b2-csi-data-dir\") pod \"csi-hostpathplugin-gzp8q\" (UID: \"47242c3c-e4f8-49f4-b508-e705396600b2\") " pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187925 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/053a3f86-3670-42c2-afe9-8b1f09ece542-config-volume\") pod \"dns-default-wlshf\" (UID: \"053a3f86-3670-42c2-afe9-8b1f09ece542\") " pod="openshift-dns/dns-default-wlshf" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.187951 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/053a3f86-3670-42c2-afe9-8b1f09ece542-metrics-tls\") pod \"dns-default-wlshf\" (UID: \"053a3f86-3670-42c2-afe9-8b1f09ece542\") " pod="openshift-dns/dns-default-wlshf" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.188017 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fc6fa0b-133a-493e-a728-3734b122fe01-serving-cert\") pod \"service-ca-operator-777779d784-jx6qd\" (UID: \"5fc6fa0b-133a-493e-a728-3734b122fe01\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jx6qd" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.188042 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c002f86-a984-4feb-b110-35aebfee516e-config\") pod \"kube-controller-manager-operator-78b949d7b-6ggzn\" (UID: \"1c002f86-a984-4feb-b110-35aebfee516e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ggzn" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.188431 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gl9h\" (UniqueName: \"kubernetes.io/projected/4af1418f-a56e-4f26-a800-3b402b7da053-kube-api-access-9gl9h\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.188489 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c1a058e6-56e0-4020-a9aa-e6537e1ccc39-signing-cabundle\") pod \"service-ca-9c57cc56f-z2dgw\" (UID: \"c1a058e6-56e0-4020-a9aa-e6537e1ccc39\") " pod="openshift-service-ca/service-ca-9c57cc56f-z2dgw" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.188576 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4af1418f-a56e-4f26-a800-3b402b7da053-trusted-ca\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.188645 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4af1418f-a56e-4f26-a800-3b402b7da053-registry-tls\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.188665 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/47242c3c-e4f8-49f4-b508-e705396600b2-mountpoint-dir\") pod \"csi-hostpathplugin-gzp8q\" (UID: \"47242c3c-e4f8-49f4-b508-e705396600b2\") " pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.188738 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e302b261-06bb-4ed0-9479-c326f26bfe43-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5w6mp\" (UID: \"e302b261-06bb-4ed0-9479-c326f26bfe43\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5w6mp" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.189477 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4af1418f-a56e-4f26-a800-3b402b7da053-ca-trust-extracted\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.191283 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4ce90879-8e80-46ec-9f0b-f972764812da-metrics-tls\") pod \"dns-operator-744455d44c-q9xpc\" (UID: \"4ce90879-8e80-46ec-9f0b-f972764812da\") " pod="openshift-dns-operator/dns-operator-744455d44c-q9xpc" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.191536 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd9f4c00-90bd-4782-8ceb-eaab6a9799c3-config\") pod \"kube-apiserver-operator-766d6c64bb-prdvb\" (UID: \"cd9f4c00-90bd-4782-8ceb-eaab6a9799c3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prdvb" Nov 24 06:57:22 crc kubenswrapper[4809]: E1124 06:57:22.192322 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:22.692305918 +0000 UTC m=+142.592897693 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.194395 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4af1418f-a56e-4f26-a800-3b402b7da053-registry-tls\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.194583 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c002f86-a984-4feb-b110-35aebfee516e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6ggzn\" (UID: \"1c002f86-a984-4feb-b110-35aebfee516e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ggzn" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.195473 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/47242c3c-e4f8-49f4-b508-e705396600b2-registration-dir\") pod \"csi-hostpathplugin-gzp8q\" (UID: \"47242c3c-e4f8-49f4-b508-e705396600b2\") " pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.195513 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5qpl\" (UniqueName: \"kubernetes.io/projected/b08de060-f5b7-4bed-a27c-89294a8827fa-kube-api-access-j5qpl\") pod \"marketplace-operator-79b997595-svxc5\" (UID: \"b08de060-f5b7-4bed-a27c-89294a8827fa\") " pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.195551 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blcz2\" (UniqueName: \"kubernetes.io/projected/47242c3c-e4f8-49f4-b508-e705396600b2-kube-api-access-blcz2\") pod \"csi-hostpathplugin-gzp8q\" (UID: \"47242c3c-e4f8-49f4-b508-e705396600b2\") " pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.197163 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4af1418f-a56e-4f26-a800-3b402b7da053-installation-pull-secrets\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.197425 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4af1418f-a56e-4f26-a800-3b402b7da053-bound-sa-token\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.197711 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1c002f86-a984-4feb-b110-35aebfee516e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6ggzn\" (UID: \"1c002f86-a984-4feb-b110-35aebfee516e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ggzn" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.197911 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/476eceb4-f8fa-49c1-9015-5e067998af80-config-volume\") pod \"collect-profiles-29399445-bkt2p\" (UID: \"476eceb4-f8fa-49c1-9015-5e067998af80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-bkt2p" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.197948 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e3ba6e1a-6434-4867-8a19-9c9a1008b688-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xq929\" (UID: \"e3ba6e1a-6434-4867-8a19-9c9a1008b688\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xq929" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.197990 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks64s\" (UniqueName: \"kubernetes.io/projected/d0b8935a-d42d-4f07-876b-67f7e1ce3d89-kube-api-access-ks64s\") pod \"machine-config-controller-84d6567774-jzwg5\" (UID: \"d0b8935a-d42d-4f07-876b-67f7e1ce3d89\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzwg5" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.198046 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3ba6e1a-6434-4867-8a19-9c9a1008b688-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xq929\" (UID: \"e3ba6e1a-6434-4867-8a19-9c9a1008b688\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xq929" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.198094 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd9f4c00-90bd-4782-8ceb-eaab6a9799c3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-prdvb\" (UID: \"cd9f4c00-90bd-4782-8ceb-eaab6a9799c3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prdvb" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.198161 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4af1418f-a56e-4f26-a800-3b402b7da053-registry-certificates\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.198180 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c07f2b54-9258-4f6d-bf9f-7f28e4331a84-images\") pod \"machine-config-operator-74547568cd-7s9hl\" (UID: \"c07f2b54-9258-4f6d-bf9f-7f28e4331a84\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7s9hl" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.198227 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c07f2b54-9258-4f6d-bf9f-7f28e4331a84-proxy-tls\") pod \"machine-config-operator-74547568cd-7s9hl\" (UID: \"c07f2b54-9258-4f6d-bf9f-7f28e4331a84\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7s9hl" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.198243 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3ba6e1a-6434-4867-8a19-9c9a1008b688-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xq929\" (UID: \"e3ba6e1a-6434-4867-8a19-9c9a1008b688\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xq929" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.198293 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3cf1e8ba-0d51-4ccf-bdf1-24959b8c5c31-certs\") pod \"machine-config-server-tk28k\" (UID: \"3cf1e8ba-0d51-4ccf-bdf1-24959b8c5c31\") " pod="openshift-machine-config-operator/machine-config-server-tk28k" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.198307 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fc6fa0b-133a-493e-a728-3734b122fe01-config\") pod \"service-ca-operator-777779d784-jx6qd\" (UID: \"5fc6fa0b-133a-493e-a728-3734b122fe01\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jx6qd" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.198336 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c07f2b54-9258-4f6d-bf9f-7f28e4331a84-auth-proxy-config\") pod \"machine-config-operator-74547568cd-7s9hl\" (UID: \"c07f2b54-9258-4f6d-bf9f-7f28e4331a84\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7s9hl" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.198353 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3cf1e8ba-0d51-4ccf-bdf1-24959b8c5c31-node-bootstrap-token\") pod \"machine-config-server-tk28k\" (UID: \"3cf1e8ba-0d51-4ccf-bdf1-24959b8c5c31\") " pod="openshift-machine-config-operator/machine-config-server-tk28k" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.199704 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3ba6e1a-6434-4867-8a19-9c9a1008b688-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xq929\" (UID: \"e3ba6e1a-6434-4867-8a19-9c9a1008b688\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xq929" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.200086 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c002f86-a984-4feb-b110-35aebfee516e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6ggzn\" (UID: \"1c002f86-a984-4feb-b110-35aebfee516e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ggzn" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.200176 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7vsp\" (UniqueName: \"kubernetes.io/projected/5fc6fa0b-133a-493e-a728-3734b122fe01-kube-api-access-h7vsp\") pod \"service-ca-operator-777779d784-jx6qd\" (UID: \"5fc6fa0b-133a-493e-a728-3734b122fe01\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jx6qd" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.200875 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4af1418f-a56e-4f26-a800-3b402b7da053-registry-certificates\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.201036 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7n7g\" (UniqueName: \"kubernetes.io/projected/50767fcc-443e-40ce-84df-2182a03691f1-kube-api-access-x7n7g\") pod \"packageserver-d55dfcdfc-r5mlr\" (UID: \"50767fcc-443e-40ce-84df-2182a03691f1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.201094 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b08de060-f5b7-4bed-a27c-89294a8827fa-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-svxc5\" (UID: \"b08de060-f5b7-4bed-a27c-89294a8827fa\") " pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.201130 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c07f2b54-9258-4f6d-bf9f-7f28e4331a84-images\") pod \"machine-config-operator-74547568cd-7s9hl\" (UID: \"c07f2b54-9258-4f6d-bf9f-7f28e4331a84\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7s9hl" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.201222 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c07f2b54-9258-4f6d-bf9f-7f28e4331a84-auth-proxy-config\") pod \"machine-config-operator-74547568cd-7s9hl\" (UID: \"c07f2b54-9258-4f6d-bf9f-7f28e4331a84\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7s9hl" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.201257 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d0b8935a-d42d-4f07-876b-67f7e1ce3d89-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jzwg5\" (UID: \"d0b8935a-d42d-4f07-876b-67f7e1ce3d89\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzwg5" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.201298 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/34fdb8d0-057e-40b9-a765-3e6a456ed218-cert\") pod \"ingress-canary-cb66b\" (UID: \"34fdb8d0-057e-40b9-a765-3e6a456ed218\") " pod="openshift-ingress-canary/ingress-canary-cb66b" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.202192 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6jp2"] Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.205209 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3ba6e1a-6434-4867-8a19-9c9a1008b688-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xq929\" (UID: \"e3ba6e1a-6434-4867-8a19-9c9a1008b688\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xq929" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.208212 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4af1418f-a56e-4f26-a800-3b402b7da053-trusted-ca\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.209030 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd9f4c00-90bd-4782-8ceb-eaab6a9799c3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-prdvb\" (UID: \"cd9f4c00-90bd-4782-8ceb-eaab6a9799c3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prdvb" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.210155 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c07f2b54-9258-4f6d-bf9f-7f28e4331a84-proxy-tls\") pod \"machine-config-operator-74547568cd-7s9hl\" (UID: \"c07f2b54-9258-4f6d-bf9f-7f28e4331a84\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7s9hl" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.224173 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrpmg\" (UniqueName: \"kubernetes.io/projected/4ce90879-8e80-46ec-9f0b-f972764812da-kube-api-access-wrpmg\") pod \"dns-operator-744455d44c-q9xpc\" (UID: \"4ce90879-8e80-46ec-9f0b-f972764812da\") " pod="openshift-dns-operator/dns-operator-744455d44c-q9xpc" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.227860 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pmsgp" Nov 24 06:57:22 crc kubenswrapper[4809]: W1124 06:57:22.237240 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e3c9b7d_5fa3_4f70_babe_522cb8dc7443.slice/crio-701686ce14ce8c04b0bcf2f610274eb8ef95eec4e450194e1da209dc9061e9ce WatchSource:0}: Error finding container 701686ce14ce8c04b0bcf2f610274eb8ef95eec4e450194e1da209dc9061e9ce: Status 404 returned error can't find the container with id 701686ce14ce8c04b0bcf2f610274eb8ef95eec4e450194e1da209dc9061e9ce Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.245370 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnc8h\" (UniqueName: \"kubernetes.io/projected/c07f2b54-9258-4f6d-bf9f-7f28e4331a84-kube-api-access-gnc8h\") pod \"machine-config-operator-74547568cd-7s9hl\" (UID: \"c07f2b54-9258-4f6d-bf9f-7f28e4331a84\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7s9hl" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.248438 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m9m27"] Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.273354 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cd9f4c00-90bd-4782-8ceb-eaab6a9799c3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-prdvb\" (UID: \"cd9f4c00-90bd-4782-8ceb-eaab6a9799c3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prdvb" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.287666 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-69dgx"] Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.293243 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl"] Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.294843 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nl2x7"] Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.296192 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gl9h\" (UniqueName: \"kubernetes.io/projected/4af1418f-a56e-4f26-a800-3b402b7da053-kube-api-access-9gl9h\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.302363 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.302531 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgqft\" (UniqueName: \"kubernetes.io/projected/3cf1e8ba-0d51-4ccf-bdf1-24959b8c5c31-kube-api-access-xgqft\") pod \"machine-config-server-tk28k\" (UID: \"3cf1e8ba-0d51-4ccf-bdf1-24959b8c5c31\") " pod="openshift-machine-config-operator/machine-config-server-tk28k" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.302556 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fc6fa0b-133a-493e-a728-3734b122fe01-serving-cert\") pod \"service-ca-operator-777779d784-jx6qd\" (UID: \"5fc6fa0b-133a-493e-a728-3734b122fe01\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jx6qd" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.302574 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/47242c3c-e4f8-49f4-b508-e705396600b2-csi-data-dir\") pod \"csi-hostpathplugin-gzp8q\" (UID: \"47242c3c-e4f8-49f4-b508-e705396600b2\") " pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.302591 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/053a3f86-3670-42c2-afe9-8b1f09ece542-config-volume\") pod \"dns-default-wlshf\" (UID: \"053a3f86-3670-42c2-afe9-8b1f09ece542\") " pod="openshift-dns/dns-default-wlshf" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.302605 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/053a3f86-3670-42c2-afe9-8b1f09ece542-metrics-tls\") pod \"dns-default-wlshf\" (UID: \"053a3f86-3670-42c2-afe9-8b1f09ece542\") " pod="openshift-dns/dns-default-wlshf" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.302622 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c1a058e6-56e0-4020-a9aa-e6537e1ccc39-signing-cabundle\") pod \"service-ca-9c57cc56f-z2dgw\" (UID: \"c1a058e6-56e0-4020-a9aa-e6537e1ccc39\") " pod="openshift-service-ca/service-ca-9c57cc56f-z2dgw" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.302648 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/47242c3c-e4f8-49f4-b508-e705396600b2-mountpoint-dir\") pod \"csi-hostpathplugin-gzp8q\" (UID: \"47242c3c-e4f8-49f4-b508-e705396600b2\") " pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.302666 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e302b261-06bb-4ed0-9479-c326f26bfe43-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5w6mp\" (UID: \"e302b261-06bb-4ed0-9479-c326f26bfe43\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5w6mp" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.302694 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/47242c3c-e4f8-49f4-b508-e705396600b2-registration-dir\") pod \"csi-hostpathplugin-gzp8q\" (UID: \"47242c3c-e4f8-49f4-b508-e705396600b2\") " pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.302710 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5qpl\" (UniqueName: \"kubernetes.io/projected/b08de060-f5b7-4bed-a27c-89294a8827fa-kube-api-access-j5qpl\") pod \"marketplace-operator-79b997595-svxc5\" (UID: \"b08de060-f5b7-4bed-a27c-89294a8827fa\") " pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.302725 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blcz2\" (UniqueName: \"kubernetes.io/projected/47242c3c-e4f8-49f4-b508-e705396600b2-kube-api-access-blcz2\") pod \"csi-hostpathplugin-gzp8q\" (UID: \"47242c3c-e4f8-49f4-b508-e705396600b2\") " pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.302767 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/476eceb4-f8fa-49c1-9015-5e067998af80-config-volume\") pod \"collect-profiles-29399445-bkt2p\" (UID: \"476eceb4-f8fa-49c1-9015-5e067998af80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-bkt2p" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.302799 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks64s\" (UniqueName: \"kubernetes.io/projected/d0b8935a-d42d-4f07-876b-67f7e1ce3d89-kube-api-access-ks64s\") pod \"machine-config-controller-84d6567774-jzwg5\" (UID: \"d0b8935a-d42d-4f07-876b-67f7e1ce3d89\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzwg5" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.302826 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3cf1e8ba-0d51-4ccf-bdf1-24959b8c5c31-node-bootstrap-token\") pod \"machine-config-server-tk28k\" (UID: \"3cf1e8ba-0d51-4ccf-bdf1-24959b8c5c31\") " pod="openshift-machine-config-operator/machine-config-server-tk28k" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.302846 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3cf1e8ba-0d51-4ccf-bdf1-24959b8c5c31-certs\") pod \"machine-config-server-tk28k\" (UID: \"3cf1e8ba-0d51-4ccf-bdf1-24959b8c5c31\") " pod="openshift-machine-config-operator/machine-config-server-tk28k" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.302865 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fc6fa0b-133a-493e-a728-3734b122fe01-config\") pod \"service-ca-operator-777779d784-jx6qd\" (UID: \"5fc6fa0b-133a-493e-a728-3734b122fe01\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jx6qd" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.302885 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7vsp\" (UniqueName: \"kubernetes.io/projected/5fc6fa0b-133a-493e-a728-3734b122fe01-kube-api-access-h7vsp\") pod \"service-ca-operator-777779d784-jx6qd\" (UID: \"5fc6fa0b-133a-493e-a728-3734b122fe01\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jx6qd" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.302915 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b08de060-f5b7-4bed-a27c-89294a8827fa-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-svxc5\" (UID: \"b08de060-f5b7-4bed-a27c-89294a8827fa\") " pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.302931 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7n7g\" (UniqueName: \"kubernetes.io/projected/50767fcc-443e-40ce-84df-2182a03691f1-kube-api-access-x7n7g\") pod \"packageserver-d55dfcdfc-r5mlr\" (UID: \"50767fcc-443e-40ce-84df-2182a03691f1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.302948 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d0b8935a-d42d-4f07-876b-67f7e1ce3d89-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jzwg5\" (UID: \"d0b8935a-d42d-4f07-876b-67f7e1ce3d89\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzwg5" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.303023 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/34fdb8d0-057e-40b9-a765-3e6a456ed218-cert\") pod \"ingress-canary-cb66b\" (UID: \"34fdb8d0-057e-40b9-a765-3e6a456ed218\") " pod="openshift-ingress-canary/ingress-canary-cb66b" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.303040 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/47242c3c-e4f8-49f4-b508-e705396600b2-socket-dir\") pod \"csi-hostpathplugin-gzp8q\" (UID: \"47242c3c-e4f8-49f4-b508-e705396600b2\") " pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.303054 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/50767fcc-443e-40ce-84df-2182a03691f1-tmpfs\") pod \"packageserver-d55dfcdfc-r5mlr\" (UID: \"50767fcc-443e-40ce-84df-2182a03691f1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.303070 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b08de060-f5b7-4bed-a27c-89294a8827fa-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-svxc5\" (UID: \"b08de060-f5b7-4bed-a27c-89294a8827fa\") " pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.303085 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e846497c-2339-4f20-b688-f2483c07d54f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pvrfw\" (UID: \"e846497c-2339-4f20-b688-f2483c07d54f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pvrfw" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.303099 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c1a058e6-56e0-4020-a9aa-e6537e1ccc39-signing-key\") pod \"service-ca-9c57cc56f-z2dgw\" (UID: \"c1a058e6-56e0-4020-a9aa-e6537e1ccc39\") " pod="openshift-service-ca/service-ca-9c57cc56f-z2dgw" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.303115 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d0b8935a-d42d-4f07-876b-67f7e1ce3d89-proxy-tls\") pod \"machine-config-controller-84d6567774-jzwg5\" (UID: \"d0b8935a-d42d-4f07-876b-67f7e1ce3d89\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzwg5" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.303132 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/50767fcc-443e-40ce-84df-2182a03691f1-apiservice-cert\") pod \"packageserver-d55dfcdfc-r5mlr\" (UID: \"50767fcc-443e-40ce-84df-2182a03691f1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.303147 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj72f\" (UniqueName: \"kubernetes.io/projected/053a3f86-3670-42c2-afe9-8b1f09ece542-kube-api-access-jj72f\") pod \"dns-default-wlshf\" (UID: \"053a3f86-3670-42c2-afe9-8b1f09ece542\") " pod="openshift-dns/dns-default-wlshf" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.303164 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/476eceb4-f8fa-49c1-9015-5e067998af80-secret-volume\") pod \"collect-profiles-29399445-bkt2p\" (UID: \"476eceb4-f8fa-49c1-9015-5e067998af80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-bkt2p" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.303181 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpgps\" (UniqueName: \"kubernetes.io/projected/e846497c-2339-4f20-b688-f2483c07d54f-kube-api-access-lpgps\") pod \"multus-admission-controller-857f4d67dd-pvrfw\" (UID: \"e846497c-2339-4f20-b688-f2483c07d54f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pvrfw" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.303195 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dm64\" (UniqueName: \"kubernetes.io/projected/c1a058e6-56e0-4020-a9aa-e6537e1ccc39-kube-api-access-4dm64\") pod \"service-ca-9c57cc56f-z2dgw\" (UID: \"c1a058e6-56e0-4020-a9aa-e6537e1ccc39\") " pod="openshift-service-ca/service-ca-9c57cc56f-z2dgw" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.303210 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhrdn\" (UniqueName: \"kubernetes.io/projected/34fdb8d0-057e-40b9-a765-3e6a456ed218-kube-api-access-vhrdn\") pod \"ingress-canary-cb66b\" (UID: \"34fdb8d0-057e-40b9-a765-3e6a456ed218\") " pod="openshift-ingress-canary/ingress-canary-cb66b" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.303227 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/50767fcc-443e-40ce-84df-2182a03691f1-webhook-cert\") pod \"packageserver-d55dfcdfc-r5mlr\" (UID: \"50767fcc-443e-40ce-84df-2182a03691f1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.303245 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jf5tz\" (UniqueName: \"kubernetes.io/projected/e302b261-06bb-4ed0-9479-c326f26bfe43-kube-api-access-jf5tz\") pod \"package-server-manager-789f6589d5-5w6mp\" (UID: \"e302b261-06bb-4ed0-9479-c326f26bfe43\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5w6mp" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.303262 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/47242c3c-e4f8-49f4-b508-e705396600b2-plugins-dir\") pod \"csi-hostpathplugin-gzp8q\" (UID: \"47242c3c-e4f8-49f4-b508-e705396600b2\") " pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.303278 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szl84\" (UniqueName: \"kubernetes.io/projected/476eceb4-f8fa-49c1-9015-5e067998af80-kube-api-access-szl84\") pod \"collect-profiles-29399445-bkt2p\" (UID: \"476eceb4-f8fa-49c1-9015-5e067998af80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-bkt2p" Nov 24 06:57:22 crc kubenswrapper[4809]: E1124 06:57:22.303342 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:22.803328176 +0000 UTC m=+142.703919881 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.304899 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b08de060-f5b7-4bed-a27c-89294a8827fa-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-svxc5\" (UID: \"b08de060-f5b7-4bed-a27c-89294a8827fa\") " pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.306193 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e846497c-2339-4f20-b688-f2483c07d54f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pvrfw\" (UID: \"e846497c-2339-4f20-b688-f2483c07d54f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pvrfw" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.306535 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/476eceb4-f8fa-49c1-9015-5e067998af80-config-volume\") pod \"collect-profiles-29399445-bkt2p\" (UID: \"476eceb4-f8fa-49c1-9015-5e067998af80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-bkt2p" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.306899 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/47242c3c-e4f8-49f4-b508-e705396600b2-registration-dir\") pod \"csi-hostpathplugin-gzp8q\" (UID: \"47242c3c-e4f8-49f4-b508-e705396600b2\") " pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.306909 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/053a3f86-3670-42c2-afe9-8b1f09ece542-config-volume\") pod \"dns-default-wlshf\" (UID: \"053a3f86-3670-42c2-afe9-8b1f09ece542\") " pod="openshift-dns/dns-default-wlshf" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.307040 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/47242c3c-e4f8-49f4-b508-e705396600b2-csi-data-dir\") pod \"csi-hostpathplugin-gzp8q\" (UID: \"47242c3c-e4f8-49f4-b508-e705396600b2\") " pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.307314 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/47242c3c-e4f8-49f4-b508-e705396600b2-plugins-dir\") pod \"csi-hostpathplugin-gzp8q\" (UID: \"47242c3c-e4f8-49f4-b508-e705396600b2\") " pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.308936 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/47242c3c-e4f8-49f4-b508-e705396600b2-mountpoint-dir\") pod \"csi-hostpathplugin-gzp8q\" (UID: \"47242c3c-e4f8-49f4-b508-e705396600b2\") " pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.309461 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c1a058e6-56e0-4020-a9aa-e6537e1ccc39-signing-key\") pod \"service-ca-9c57cc56f-z2dgw\" (UID: \"c1a058e6-56e0-4020-a9aa-e6537e1ccc39\") " pod="openshift-service-ca/service-ca-9c57cc56f-z2dgw" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.310724 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d0b8935a-d42d-4f07-876b-67f7e1ce3d89-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jzwg5\" (UID: \"d0b8935a-d42d-4f07-876b-67f7e1ce3d89\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzwg5" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.311270 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c1a058e6-56e0-4020-a9aa-e6537e1ccc39-signing-cabundle\") pod \"service-ca-9c57cc56f-z2dgw\" (UID: \"c1a058e6-56e0-4020-a9aa-e6537e1ccc39\") " pod="openshift-service-ca/service-ca-9c57cc56f-z2dgw" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.313623 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3cf1e8ba-0d51-4ccf-bdf1-24959b8c5c31-node-bootstrap-token\") pod \"machine-config-server-tk28k\" (UID: \"3cf1e8ba-0d51-4ccf-bdf1-24959b8c5c31\") " pod="openshift-machine-config-operator/machine-config-server-tk28k" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.316218 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fc6fa0b-133a-493e-a728-3734b122fe01-config\") pod \"service-ca-operator-777779d784-jx6qd\" (UID: \"5fc6fa0b-133a-493e-a728-3734b122fe01\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jx6qd" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.316218 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fc6fa0b-133a-493e-a728-3734b122fe01-serving-cert\") pod \"service-ca-operator-777779d784-jx6qd\" (UID: \"5fc6fa0b-133a-493e-a728-3734b122fe01\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jx6qd" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.317383 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/50767fcc-443e-40ce-84df-2182a03691f1-tmpfs\") pod \"packageserver-d55dfcdfc-r5mlr\" (UID: \"50767fcc-443e-40ce-84df-2182a03691f1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.317460 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/47242c3c-e4f8-49f4-b508-e705396600b2-socket-dir\") pod \"csi-hostpathplugin-gzp8q\" (UID: \"47242c3c-e4f8-49f4-b508-e705396600b2\") " pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.317871 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e302b261-06bb-4ed0-9479-c326f26bfe43-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5w6mp\" (UID: \"e302b261-06bb-4ed0-9479-c326f26bfe43\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5w6mp" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.318456 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d0b8935a-d42d-4f07-876b-67f7e1ce3d89-proxy-tls\") pod \"machine-config-controller-84d6567774-jzwg5\" (UID: \"d0b8935a-d42d-4f07-876b-67f7e1ce3d89\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzwg5" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.318629 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/50767fcc-443e-40ce-84df-2182a03691f1-apiservice-cert\") pod \"packageserver-d55dfcdfc-r5mlr\" (UID: \"50767fcc-443e-40ce-84df-2182a03691f1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.319376 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3cf1e8ba-0d51-4ccf-bdf1-24959b8c5c31-certs\") pod \"machine-config-server-tk28k\" (UID: \"3cf1e8ba-0d51-4ccf-bdf1-24959b8c5c31\") " pod="openshift-machine-config-operator/machine-config-server-tk28k" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.320310 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/50767fcc-443e-40ce-84df-2182a03691f1-webhook-cert\") pod \"packageserver-d55dfcdfc-r5mlr\" (UID: \"50767fcc-443e-40ce-84df-2182a03691f1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.324167 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/34fdb8d0-057e-40b9-a765-3e6a456ed218-cert\") pod \"ingress-canary-cb66b\" (UID: \"34fdb8d0-057e-40b9-a765-3e6a456ed218\") " pod="openshift-ingress-canary/ingress-canary-cb66b" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.328980 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b08de060-f5b7-4bed-a27c-89294a8827fa-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-svxc5\" (UID: \"b08de060-f5b7-4bed-a27c-89294a8827fa\") " pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.330094 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/476eceb4-f8fa-49c1-9015-5e067998af80-secret-volume\") pod \"collect-profiles-29399445-bkt2p\" (UID: \"476eceb4-f8fa-49c1-9015-5e067998af80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-bkt2p" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.331304 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/053a3f86-3670-42c2-afe9-8b1f09ece542-metrics-tls\") pod \"dns-default-wlshf\" (UID: \"053a3f86-3670-42c2-afe9-8b1f09ece542\") " pod="openshift-dns/dns-default-wlshf" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.334326 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rffdv" event={"ID":"dd0770b5-d9e3-48ae-9ce0-58bcd6a47207","Type":"ContainerStarted","Data":"5f6952478fbaacbc5bf5a09fcd0e8534024b073f8f78f7c83c2d480b38e0e71d"} Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.335817 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1c002f86-a984-4feb-b110-35aebfee516e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6ggzn\" (UID: \"1c002f86-a984-4feb-b110-35aebfee516e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ggzn" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.337905 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cj64v" event={"ID":"25780238-d742-4461-af80-a4808c4dd40b","Type":"ContainerStarted","Data":"280d06613fc9f4850e870c44abc75e4b6c531e64270da6931851b16245c72dea"} Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.345130 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q8lpg" event={"ID":"1bace469-be4b-48d6-9152-d65a4b762f5d","Type":"ContainerStarted","Data":"16aa226c045a8ba437a433f1387eed48ccc0c02adb1bc627309d75b842e41c38"} Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.349555 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e3ba6e1a-6434-4867-8a19-9c9a1008b688-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xq929\" (UID: \"e3ba6e1a-6434-4867-8a19-9c9a1008b688\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xq929" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.350317 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-lkdj2" event={"ID":"b6237cc7-ba93-45a6-b560-d7916abf71f7","Type":"ContainerStarted","Data":"1588b9ef20d3dff755bfdf824fb2fda2c0bec4f1eeb2f3e108adfd7d4bf3bace"} Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.365483 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-88wkc" event={"ID":"66ef6324-33df-4e2a-92dd-36a3f5f7995d","Type":"ContainerStarted","Data":"c037df14e558b91faa4a32e37d6b23118454073bc55dc7550797fe1b8adbae9f"} Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.367133 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-lxf68" event={"ID":"df85fd23-9a76-41a3-b17d-7d06b5bfda87","Type":"ContainerStarted","Data":"aa28271282722b226d613866756086abbd31320beba7b3e99c8e0c73fc94e537"} Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.368865 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6jp2" event={"ID":"4e3c9b7d-5fa3-4f70-babe-522cb8dc7443","Type":"ContainerStarted","Data":"701686ce14ce8c04b0bcf2f610274eb8ef95eec4e450194e1da209dc9061e9ce"} Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.370383 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4af1418f-a56e-4f26-a800-3b402b7da053-bound-sa-token\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.371429 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" event={"ID":"04b0f650-8f86-4d8e-9fbc-860516a0f3a7","Type":"ContainerStarted","Data":"8d62e7affaa79a8150fd8340f70c00aedcf1066b5200cf3af835ad3fe123f223"} Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.378304 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-rpd99" event={"ID":"8f1b5104-f6eb-4e51-9339-e63f4429f02c","Type":"ContainerStarted","Data":"4256b0c8410e29cb0cf2ba36adb67db085f15a8a9769728c0c8115daa071752c"} Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.378341 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-rpd99" event={"ID":"8f1b5104-f6eb-4e51-9339-e63f4429f02c","Type":"ContainerStarted","Data":"dcb6da8af86db4e2af72acebf3485dcd4a7c31ebf467267c4c21d2a8ac2e3a3d"} Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.378757 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-rpd99" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.384948 4809 patch_prober.go:28] interesting pod/console-operator-58897d9998-rpd99 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.31:8443/readyz\": dial tcp 10.217.0.31:8443: connect: connection refused" start-of-body= Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.385019 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-rpd99" podUID="8f1b5104-f6eb-4e51-9339-e63f4429f02c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.31:8443/readyz\": dial tcp 10.217.0.31:8443: connect: connection refused" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.402778 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" event={"ID":"f3ac0673-5554-4ac6-a16e-b08631fdbf94","Type":"ContainerStarted","Data":"e33de40b31ff2c23d7d9a91492598171a4d75feb1523db57b411336a7596464f"} Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.403019 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" event={"ID":"f3ac0673-5554-4ac6-a16e-b08631fdbf94","Type":"ContainerStarted","Data":"cc23e664312807f16fb1c5601b42d4544c2e8d9fb7a5acf7afc28219cfafbe3d"} Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.403033 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.404070 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.404079 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5qpl\" (UniqueName: \"kubernetes.io/projected/b08de060-f5b7-4bed-a27c-89294a8827fa-kube-api-access-j5qpl\") pod \"marketplace-operator-79b997595-svxc5\" (UID: \"b08de060-f5b7-4bed-a27c-89294a8827fa\") " pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" Nov 24 06:57:22 crc kubenswrapper[4809]: E1124 06:57:22.404330 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:22.904316642 +0000 UTC m=+142.804908417 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.406359 4809 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-92k7t container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.406389 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" podUID="f3ac0673-5554-4ac6-a16e-b08631fdbf94" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.406499 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m9m27" event={"ID":"cf7f6885-ff5e-44df-ac63-f00fa3ef63a6","Type":"ContainerStarted","Data":"de95b0bc6df19b6e4311598ab5554fa8c8a3732a27c486214acaf15f96af3b1b"} Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.414286 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-r7k56" event={"ID":"9eab3672-3dd0-4af1-b96d-e8c5e99bb840","Type":"ContainerStarted","Data":"e2728fc554c6b89ad83f534cf56eac2b9f022b2ed2cdfcd1c3318dc64c2d4628"} Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.440812 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgqft\" (UniqueName: \"kubernetes.io/projected/3cf1e8ba-0d51-4ccf-bdf1-24959b8c5c31-kube-api-access-xgqft\") pod \"machine-config-server-tk28k\" (UID: \"3cf1e8ba-0d51-4ccf-bdf1-24959b8c5c31\") " pod="openshift-machine-config-operator/machine-config-server-tk28k" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.445392 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ggzn" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.467512 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpgps\" (UniqueName: \"kubernetes.io/projected/e846497c-2339-4f20-b688-f2483c07d54f-kube-api-access-lpgps\") pod \"multus-admission-controller-857f4d67dd-pvrfw\" (UID: \"e846497c-2339-4f20-b688-f2483c07d54f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pvrfw" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.476283 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xq929" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.478273 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blcz2\" (UniqueName: \"kubernetes.io/projected/47242c3c-e4f8-49f4-b508-e705396600b2-kube-api-access-blcz2\") pod \"csi-hostpathplugin-gzp8q\" (UID: \"47242c3c-e4f8-49f4-b508-e705396600b2\") " pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.484588 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-54rmc"] Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.487639 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prdvb" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.494912 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mqs58"] Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.496818 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dm64\" (UniqueName: \"kubernetes.io/projected/c1a058e6-56e0-4020-a9aa-e6537e1ccc39-kube-api-access-4dm64\") pod \"service-ca-9c57cc56f-z2dgw\" (UID: \"c1a058e6-56e0-4020-a9aa-e6537e1ccc39\") " pod="openshift-service-ca/service-ca-9c57cc56f-z2dgw" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.504335 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7s9hl" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.505021 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:22 crc kubenswrapper[4809]: E1124 06:57:22.506376 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:23.006355117 +0000 UTC m=+142.906946822 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.506520 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-56j57"] Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.526843 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-q9xpc" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.529899 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhrdn\" (UniqueName: \"kubernetes.io/projected/34fdb8d0-057e-40b9-a765-3e6a456ed218-kube-api-access-vhrdn\") pod \"ingress-canary-cb66b\" (UID: \"34fdb8d0-057e-40b9-a765-3e6a456ed218\") " pod="openshift-ingress-canary/ingress-canary-cb66b" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.540828 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks64s\" (UniqueName: \"kubernetes.io/projected/d0b8935a-d42d-4f07-876b-67f7e1ce3d89-kube-api-access-ks64s\") pod \"machine-config-controller-84d6567774-jzwg5\" (UID: \"d0b8935a-d42d-4f07-876b-67f7e1ce3d89\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzwg5" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.547093 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jf5tz\" (UniqueName: \"kubernetes.io/projected/e302b261-06bb-4ed0-9479-c326f26bfe43-kube-api-access-jf5tz\") pod \"package-server-manager-789f6589d5-5w6mp\" (UID: \"e302b261-06bb-4ed0-9479-c326f26bfe43\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5w6mp" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.548661 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-72wvs"] Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.552907 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzwg5" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.565099 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-z2dgw" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.567307 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj72f\" (UniqueName: \"kubernetes.io/projected/053a3f86-3670-42c2-afe9-8b1f09ece542-kube-api-access-jj72f\") pod \"dns-default-wlshf\" (UID: \"053a3f86-3670-42c2-afe9-8b1f09ece542\") " pod="openshift-dns/dns-default-wlshf" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.569665 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5w6mp" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.585569 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7n7g\" (UniqueName: \"kubernetes.io/projected/50767fcc-443e-40ce-84df-2182a03691f1-kube-api-access-x7n7g\") pod \"packageserver-d55dfcdfc-r5mlr\" (UID: \"50767fcc-443e-40ce-84df-2182a03691f1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr" Nov 24 06:57:22 crc kubenswrapper[4809]: W1124 06:57:22.590332 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeca6947b_dd5e_449d_a8c5_05277c556bd1.slice/crio-9688fc61f14fd7319d238932262adba775504bf4fe3b4a534934cd87bbd98bc9 WatchSource:0}: Error finding container 9688fc61f14fd7319d238932262adba775504bf4fe3b4a534934cd87bbd98bc9: Status 404 returned error can't find the container with id 9688fc61f14fd7319d238932262adba775504bf4fe3b4a534934cd87bbd98bc9 Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.590469 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-tk28k" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.599147 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zhgn2"] Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.602600 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-pvrfw" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.606099 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: E1124 06:57:22.606472 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:23.106453065 +0000 UTC m=+143.007044770 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.611214 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szl84\" (UniqueName: \"kubernetes.io/projected/476eceb4-f8fa-49c1-9015-5e067998af80-kube-api-access-szl84\") pod \"collect-profiles-29399445-bkt2p\" (UID: \"476eceb4-f8fa-49c1-9015-5e067998af80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-bkt2p" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.616497 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" Nov 24 06:57:22 crc kubenswrapper[4809]: W1124 06:57:22.617096 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1a1cfe0_beab_41d4_a478_668f5e7386f9.slice/crio-a59a49c058f4f0dbb496aea42b1fec54e8ba8d004c57e75830a1720d226afb17 WatchSource:0}: Error finding container a59a49c058f4f0dbb496aea42b1fec54e8ba8d004c57e75830a1720d226afb17: Status 404 returned error can't find the container with id a59a49c058f4f0dbb496aea42b1fec54e8ba8d004c57e75830a1720d226afb17 Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.628827 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7vsp\" (UniqueName: \"kubernetes.io/projected/5fc6fa0b-133a-493e-a728-3734b122fe01-kube-api-access-h7vsp\") pod \"service-ca-operator-777779d784-jx6qd\" (UID: \"5fc6fa0b-133a-493e-a728-3734b122fe01\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jx6qd" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.647762 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.650960 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-cb66b" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.662845 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pmsgp"] Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.700451 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-wlshf" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.707384 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:22 crc kubenswrapper[4809]: E1124 06:57:22.707632 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:23.207605775 +0000 UTC m=+143.108197480 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.707717 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.707868 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-8r5vw"] Nov 24 06:57:22 crc kubenswrapper[4809]: E1124 06:57:22.708133 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:23.20812617 +0000 UTC m=+143.108717875 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.751718 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mmdhq"] Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.765747 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xfdvj"] Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.808711 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:22 crc kubenswrapper[4809]: E1124 06:57:22.809291 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:23.309272269 +0000 UTC m=+143.209863974 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:22 crc kubenswrapper[4809]: W1124 06:57:22.836157 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8dd70f5f_be62_4eda_8cb0_c197ca2bb5a4.slice/crio-e84aa4c2d517e2e30aeb806295566a7c48179e02f9ef4e8ee76dbc392a62ea75 WatchSource:0}: Error finding container e84aa4c2d517e2e30aeb806295566a7c48179e02f9ef4e8ee76dbc392a62ea75: Status 404 returned error can't find the container with id e84aa4c2d517e2e30aeb806295566a7c48179e02f9ef4e8ee76dbc392a62ea75 Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.877892 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.884044 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-jx6qd" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.915595 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-bkt2p" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.916076 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:22 crc kubenswrapper[4809]: E1124 06:57:22.917148 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:23.417133845 +0000 UTC m=+143.317725550 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.917824 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" podStartSLOduration=122.917805025 podStartE2EDuration="2m2.917805025s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:22.909736578 +0000 UTC m=+142.810328303" watchObservedRunningTime="2025-11-24 06:57:22.917805025 +0000 UTC m=+142.818396730" Nov 24 06:57:22 crc kubenswrapper[4809]: I1124 06:57:22.919573 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ggzn"] Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.017339 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:23 crc kubenswrapper[4809]: E1124 06:57:23.017745 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:23.517729698 +0000 UTC m=+143.418321403 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.042027 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prdvb"] Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.044071 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xq929"] Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.119246 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:23 crc kubenswrapper[4809]: E1124 06:57:23.119581 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:23.619565777 +0000 UTC m=+143.520157482 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:23 crc kubenswrapper[4809]: W1124 06:57:23.124169 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd9f4c00_90bd_4782_8ceb_eaab6a9799c3.slice/crio-8578431d5a95d0b8cca60aad60f6ebb33ef28cf96fcda66d6765f99044a0cb88 WatchSource:0}: Error finding container 8578431d5a95d0b8cca60aad60f6ebb33ef28cf96fcda66d6765f99044a0cb88: Status 404 returned error can't find the container with id 8578431d5a95d0b8cca60aad60f6ebb33ef28cf96fcda66d6765f99044a0cb88 Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.143093 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-7s9hl"] Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.220019 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:23 crc kubenswrapper[4809]: E1124 06:57:23.220259 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:23.720233953 +0000 UTC m=+143.620825648 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.220623 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:23 crc kubenswrapper[4809]: E1124 06:57:23.229612 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:23.729590337 +0000 UTC m=+143.630182042 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.322189 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:23 crc kubenswrapper[4809]: E1124 06:57:23.322674 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:23.822656989 +0000 UTC m=+143.723248694 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.336179 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pvrfw"] Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.423766 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:23 crc kubenswrapper[4809]: E1124 06:57:23.424369 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:23.924358075 +0000 UTC m=+143.824949780 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:23 crc kubenswrapper[4809]: W1124 06:57:23.428734 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode846497c_2339_4f20_b688_f2483c07d54f.slice/crio-2fe2f8e2940300eb7aaf29c5386eb92db6e38430ae015307510c40a52dfe4c86 WatchSource:0}: Error finding container 2fe2f8e2940300eb7aaf29c5386eb92db6e38430ae015307510c40a52dfe4c86: Status 404 returned error can't find the container with id 2fe2f8e2940300eb7aaf29c5386eb92db6e38430ae015307510c40a52dfe4c86 Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.431735 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jzwg5"] Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.464484 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pmsgp" event={"ID":"8b246b95-be2e-4c90-8cf0-f92c612bfbc1","Type":"ContainerStarted","Data":"56541b7226f4ee81600024375bf56b91cacc2543245b7726632d96605047b26f"} Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.479782 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-mmdhq" event={"ID":"8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4","Type":"ContainerStarted","Data":"e84aa4c2d517e2e30aeb806295566a7c48179e02f9ef4e8ee76dbc392a62ea75"} Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.486843 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-lkdj2" event={"ID":"b6237cc7-ba93-45a6-b560-d7916abf71f7","Type":"ContainerStarted","Data":"f84951763724197840b80fadfe5a95cc4d6501141c8b413805cd03162e583893"} Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.487883 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-lkdj2" Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.497652 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-lkdj2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.497707 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lkdj2" podUID="b6237cc7-ba93-45a6-b560-d7916abf71f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.527396 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:23 crc kubenswrapper[4809]: E1124 06:57:23.527876 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:24.027858813 +0000 UTC m=+143.928450518 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.528100 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:23 crc kubenswrapper[4809]: E1124 06:57:23.529268 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:24.029254683 +0000 UTC m=+143.929846388 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.550720 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rffdv" event={"ID":"dd0770b5-d9e3-48ae-9ce0-58bcd6a47207","Type":"ContainerStarted","Data":"939cbc7d6783f2b09179e85b8c7d68fb2114702f58498216a6fa2ee64d1ec1bc"} Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.594787 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m9m27" event={"ID":"cf7f6885-ff5e-44df-ac63-f00fa3ef63a6","Type":"ContainerStarted","Data":"02a892848f7a2224757bf9ae7072160e523a6ecdba6e92db610bce718081e900"} Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.604636 4809 generic.go:334] "Generic (PLEG): container finished" podID="66ef6324-33df-4e2a-92dd-36a3f5f7995d" containerID="49dbe29d05cd2e2375e1dd3f692bfca720f3d2967b7f49841f34bce11a9bdef5" exitCode=0 Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.604703 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-88wkc" event={"ID":"66ef6324-33df-4e2a-92dd-36a3f5f7995d","Type":"ContainerDied","Data":"49dbe29d05cd2e2375e1dd3f692bfca720f3d2967b7f49841f34bce11a9bdef5"} Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.626110 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8r5vw" event={"ID":"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884","Type":"ContainerStarted","Data":"55a4bf0e862c2ea68500c04a7e69e8261fc4d00c00343076a1cef7ecaaa25189"} Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.628386 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:23 crc kubenswrapper[4809]: E1124 06:57:23.629829 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:24.129814005 +0000 UTC m=+144.030405710 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.639220 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" event={"ID":"358682a8-6b18-49ee-bd7d-d9eab2b22ee0","Type":"ContainerStarted","Data":"fc4e9365dbecfca96183bb8ac808a45f2fae8044e741583d2c8e1bcd913f9d74"} Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.639259 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" event={"ID":"358682a8-6b18-49ee-bd7d-d9eab2b22ee0","Type":"ContainerStarted","Data":"713f6b38bee99b89dc7907e191f636f646dda0aa80d016e0fc10e719f3843c75"} Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.639811 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.682003 4809 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-p86rl container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.682343 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" podUID="358682a8-6b18-49ee-bd7d-d9eab2b22ee0" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.731025 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mqs58" event={"ID":"7bf55845-293a-497e-87f1-eb5c36670105","Type":"ContainerStarted","Data":"af37620723d5e46ca4ab1606b7a95d84f5fc7a7b9d2a2543fb08e86e79319db4"} Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.732116 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:23 crc kubenswrapper[4809]: E1124 06:57:23.735538 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:24.235521599 +0000 UTC m=+144.136113304 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.773089 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-gzp8q"] Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.780945 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-lxf68" event={"ID":"df85fd23-9a76-41a3-b17d-7d06b5bfda87","Type":"ContainerStarted","Data":"53a005d121c02b3ab52d78550d46de9314783d03f6b9772672ad276b8ac92cfb"} Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.805169 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q9xpc"] Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.832048 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-z2dgw"] Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.832876 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:23 crc kubenswrapper[4809]: E1124 06:57:23.833249 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:24.333233877 +0000 UTC m=+144.233825582 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.833524 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399445-bkt2p"] Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.845511 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-svxc5"] Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.868343 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-wlshf"] Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.870725 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xfdvj" event={"ID":"3e14d56e-43b3-40ec-809f-371e8ab2f628","Type":"ContainerStarted","Data":"eda7fb1a9ff2cf46d8b6d64a29a16a0961f6d7a98e2db54429743cacb9114ccf"} Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.870772 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xfdvj" event={"ID":"3e14d56e-43b3-40ec-809f-371e8ab2f628","Type":"ContainerStarted","Data":"edce1039f4ae9b62662fea26f7e92d318a7a862e2fb4259ab01e32a57c9d56de"} Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.888648 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6jp2" event={"ID":"4e3c9b7d-5fa3-4f70-babe-522cb8dc7443","Type":"ContainerStarted","Data":"2fbfb2b2d3bb476f46c3f82980337f6c23fdbd6a64d914799acc5fa3a12cdd5e"} Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.909803 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-rpd99" podStartSLOduration=123.909788264 podStartE2EDuration="2m3.909788264s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:23.909323171 +0000 UTC m=+143.809914876" watchObservedRunningTime="2025-11-24 06:57:23.909788264 +0000 UTC m=+143.810379969" Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.917703 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-tk28k" event={"ID":"3cf1e8ba-0d51-4ccf-bdf1-24959b8c5c31","Type":"ContainerStarted","Data":"d2f85668f48820f09d9b0e9a708e4eeccce3fdec00dd625c16037ae434fccdcf"} Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.926189 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5w6mp"] Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.928407 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-cb66b"] Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.938077 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.940854 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xq929" event={"ID":"e3ba6e1a-6434-4867-8a19-9c9a1008b688","Type":"ContainerStarted","Data":"d37d731d4599d0d03764815cfc5daf73de003d86263095d8eac62158699e93f6"} Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.942246 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7s9hl" event={"ID":"c07f2b54-9258-4f6d-bf9f-7f28e4331a84","Type":"ContainerStarted","Data":"fdacb69ae902cb6c25e73e5ee569982cf82df6f86205af9f292c1dabcd747cee"} Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.950757 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q8lpg" event={"ID":"1bace469-be4b-48d6-9152-d65a4b762f5d","Type":"ContainerStarted","Data":"024f3f9516495e6f8f8090a949da7116fadc225cff13bd3b769b8ced99125777"} Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.950796 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q8lpg" event={"ID":"1bace469-be4b-48d6-9152-d65a4b762f5d","Type":"ContainerStarted","Data":"096de9850d9550f6849f6c50c043ecffae0cc060db6d61996033821d409ba4aa"} Nov 24 06:57:23 crc kubenswrapper[4809]: E1124 06:57:23.953212 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:24.439110464 +0000 UTC m=+144.339702169 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.955899 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zhgn2" event={"ID":"c3b64526-f2a8-4141-b04c-65b65be09ccc","Type":"ContainerStarted","Data":"d293f376d80a853b9a67127943d5a1d06fdc336089c24c9909debb2fd5951a1b"} Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.960054 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-r7k56" event={"ID":"9eab3672-3dd0-4af1-b96d-e8c5e99bb840","Type":"ContainerStarted","Data":"fbd8461b373f455af5de2e73058d5d8afb975655fa47cd1e4bf0381955b0a688"} Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.963688 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ggzn" event={"ID":"1c002f86-a984-4feb-b110-35aebfee516e","Type":"ContainerStarted","Data":"1eb9815896039c3abc4f30a76bc68cc874799133cb7195ab713a26d7b82e8267"} Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.992391 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-72wvs" event={"ID":"f1a1cfe0-beab-41d4-a478-668f5e7386f9","Type":"ContainerStarted","Data":"a59a49c058f4f0dbb496aea42b1fec54e8ba8d004c57e75830a1720d226afb17"} Nov 24 06:57:23 crc kubenswrapper[4809]: I1124 06:57:23.993357 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-72wvs" Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.022254 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-54rmc" event={"ID":"96284acd-e6e2-467c-8c70-005ca153dda3","Type":"ContainerStarted","Data":"f92de7dbbe52aa65fb17587174e4e705af85db2a63afe320963e79b520d7f7cf"} Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.022308 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-54rmc" event={"ID":"96284acd-e6e2-467c-8c70-005ca153dda3","Type":"ContainerStarted","Data":"89a282546c64e8b36753cf5667fcf9dfafb8a892e4cb36cd2415144d1249706e"} Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.030099 4809 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-72wvs container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.030151 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-72wvs" podUID="f1a1cfe0-beab-41d4-a478-668f5e7386f9" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.038542 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:24 crc kubenswrapper[4809]: E1124 06:57:24.043657 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:24.543637273 +0000 UTC m=+144.444228978 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.060316 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nl2x7" event={"ID":"fe7c3c56-9f9c-4120-9a89-f840aff6aac8","Type":"ContainerStarted","Data":"d758aa4973d8e6e42594b4a6fed43396fc0a58f30321bba8dc661c3b9d4ced10"} Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.060365 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nl2x7" event={"ID":"fe7c3c56-9f9c-4120-9a89-f840aff6aac8","Type":"ContainerStarted","Data":"45a89d63cd78c24f93f1f042626363d74951db3264b5c50e5b891ddeacbe90df"} Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.121200 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cj64v" event={"ID":"25780238-d742-4461-af80-a4808c4dd40b","Type":"ContainerStarted","Data":"54868ca586272bac1201f017c8dc957a533b4b04245d935fc84d05a7b4f152ac"} Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.121493 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cj64v" event={"ID":"25780238-d742-4461-af80-a4808c4dd40b","Type":"ContainerStarted","Data":"f10508d550f3565ed3342a96734694fb525dde2cbac5b069d8e4b6bcf8199231"} Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.124148 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-r7k56" Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.140942 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:24 crc kubenswrapper[4809]: E1124 06:57:24.141645 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:24.64162858 +0000 UTC m=+144.542220275 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.142336 4809 patch_prober.go:28] interesting pod/router-default-5444994796-r7k56 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:57:24 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Nov 24 06:57:24 crc kubenswrapper[4809]: [+]process-running ok Nov 24 06:57:24 crc kubenswrapper[4809]: healthz check failed Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.142370 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r7k56" podUID="9eab3672-3dd0-4af1-b96d-e8c5e99bb840" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.156324 4809 generic.go:334] "Generic (PLEG): container finished" podID="04b0f650-8f86-4d8e-9fbc-860516a0f3a7" containerID="c708d643c87ff69f06574471bcac45d0c9912c38ce3ce238e590498dc94597fc" exitCode=0 Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.156440 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" event={"ID":"04b0f650-8f86-4d8e-9fbc-860516a0f3a7","Type":"ContainerDied","Data":"c708d643c87ff69f06574471bcac45d0c9912c38ce3ce238e590498dc94597fc"} Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.170251 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-lxf68" podStartSLOduration=124.170232169 podStartE2EDuration="2m4.170232169s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:24.157488715 +0000 UTC m=+144.058080420" watchObservedRunningTime="2025-11-24 06:57:24.170232169 +0000 UTC m=+144.070823874" Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.203781 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-jx6qd"] Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.209152 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr"] Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.217381 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" event={"ID":"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf","Type":"ContainerStarted","Data":"473a63602d48c8f4db20c57684dd8ed5918e2686427a121d8c5c7f87130c5d2b"} Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.217427 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" event={"ID":"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf","Type":"ContainerStarted","Data":"a3277eb73eeb2de4365bdcb3f77cfed5c6cd44a3052bfc77b835dcfc2e3bd4ed"} Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.218203 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.223951 4809 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-69dgx container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.224006 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" podUID="51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.226542 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prdvb" event={"ID":"cd9f4c00-90bd-4782-8ceb-eaab6a9799c3","Type":"ContainerStarted","Data":"8578431d5a95d0b8cca60aad60f6ebb33ef28cf96fcda66d6765f99044a0cb88"} Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.241870 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:24 crc kubenswrapper[4809]: E1124 06:57:24.242058 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:24.742039218 +0000 UTC m=+144.642630923 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.242111 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:24 crc kubenswrapper[4809]: E1124 06:57:24.243774 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:24.743762198 +0000 UTC m=+144.644353903 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.264291 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m9m27" podStartSLOduration=124.264210768 podStartE2EDuration="2m4.264210768s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:24.238247956 +0000 UTC m=+144.138839661" watchObservedRunningTime="2025-11-24 06:57:24.264210768 +0000 UTC m=+144.164802473" Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.264649 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-tk28k" podStartSLOduration=5.264641801 podStartE2EDuration="5.264641801s" podCreationTimestamp="2025-11-24 06:57:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:24.263132566 +0000 UTC m=+144.163724271" watchObservedRunningTime="2025-11-24 06:57:24.264641801 +0000 UTC m=+144.165233506" Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.277040 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-56j57" event={"ID":"eca6947b-dd5e-449d-a8c5-05277c556bd1","Type":"ContainerStarted","Data":"9688fc61f14fd7319d238932262adba775504bf4fe3b4a534934cd87bbd98bc9"} Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.301945 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-rpd99" Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.305457 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.342767 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:24 crc kubenswrapper[4809]: E1124 06:57:24.343521 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:24.843493855 +0000 UTC m=+144.744085621 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.449066 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.449177 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" podStartSLOduration=124.449160227 podStartE2EDuration="2m4.449160227s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:24.406638789 +0000 UTC m=+144.307230494" watchObservedRunningTime="2025-11-24 06:57:24.449160227 +0000 UTC m=+144.349751932" Nov 24 06:57:24 crc kubenswrapper[4809]: E1124 06:57:24.449415 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:24.949403615 +0000 UTC m=+144.849995320 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.450215 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q8lpg" podStartSLOduration=124.450209558 podStartE2EDuration="2m4.450209558s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:24.447944771 +0000 UTC m=+144.348536466" watchObservedRunningTime="2025-11-24 06:57:24.450209558 +0000 UTC m=+144.350801263" Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.486252 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6jp2" podStartSLOduration=124.486234816 podStartE2EDuration="2m4.486234816s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:24.484275928 +0000 UTC m=+144.384867623" watchObservedRunningTime="2025-11-24 06:57:24.486234816 +0000 UTC m=+144.386826521" Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.549867 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:24 crc kubenswrapper[4809]: E1124 06:57:24.550312 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:25.050294336 +0000 UTC m=+144.950886041 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.606262 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-72wvs" podStartSLOduration=124.606243709 podStartE2EDuration="2m4.606243709s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:24.574431174 +0000 UTC m=+144.475022899" watchObservedRunningTime="2025-11-24 06:57:24.606243709 +0000 UTC m=+144.506835414" Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.653658 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:24 crc kubenswrapper[4809]: E1124 06:57:24.654229 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:25.154218017 +0000 UTC m=+145.054809722 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.662281 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nl2x7" podStartSLOduration=124.662261843 podStartE2EDuration="2m4.662261843s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:24.660829951 +0000 UTC m=+144.561421656" watchObservedRunningTime="2025-11-24 06:57:24.662261843 +0000 UTC m=+144.562853548" Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.755089 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:24 crc kubenswrapper[4809]: E1124 06:57:24.755263 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:25.255219222 +0000 UTC m=+145.155810927 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.755477 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:24 crc kubenswrapper[4809]: E1124 06:57:24.755787 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:25.255779928 +0000 UTC m=+145.156371633 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.841150 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-lkdj2" podStartSLOduration=124.841136843 podStartE2EDuration="2m4.841136843s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:24.840637009 +0000 UTC m=+144.741228714" watchObservedRunningTime="2025-11-24 06:57:24.841136843 +0000 UTC m=+144.741728548" Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.857246 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:24 crc kubenswrapper[4809]: E1124 06:57:24.857670 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:25.357655058 +0000 UTC m=+145.258246763 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.897174 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rffdv" podStartSLOduration=124.897157458 podStartE2EDuration="2m4.897157458s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:24.896917761 +0000 UTC m=+144.797509466" watchObservedRunningTime="2025-11-24 06:57:24.897157458 +0000 UTC m=+144.797749163" Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.950756 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-r7k56" podStartSLOduration=124.950738871 podStartE2EDuration="2m4.950738871s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:24.950190304 +0000 UTC m=+144.850782019" watchObservedRunningTime="2025-11-24 06:57:24.950738871 +0000 UTC m=+144.851330576" Nov 24 06:57:24 crc kubenswrapper[4809]: I1124 06:57:24.962666 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:24 crc kubenswrapper[4809]: E1124 06:57:24.963198 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:25.463184906 +0000 UTC m=+145.363776611 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.004946 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xfdvj" podStartSLOduration=125.004924512 podStartE2EDuration="2m5.004924512s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:24.99669947 +0000 UTC m=+144.897291165" watchObservedRunningTime="2025-11-24 06:57:25.004924512 +0000 UTC m=+144.905516217" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.064271 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:25 crc kubenswrapper[4809]: E1124 06:57:25.064589 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:25.564575402 +0000 UTC m=+145.465167107 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.112513 4809 patch_prober.go:28] interesting pod/router-default-5444994796-r7k56 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:57:25 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Nov 24 06:57:25 crc kubenswrapper[4809]: [+]process-running ok Nov 24 06:57:25 crc kubenswrapper[4809]: healthz check failed Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.112575 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r7k56" podUID="9eab3672-3dd0-4af1-b96d-e8c5e99bb840" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.166909 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:25 crc kubenswrapper[4809]: E1124 06:57:25.167344 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:25.667330489 +0000 UTC m=+145.567922194 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.269984 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:25 crc kubenswrapper[4809]: E1124 06:57:25.270577 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:25.770562329 +0000 UTC m=+145.671154034 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.272466 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cj64v" podStartSLOduration=125.272449314 podStartE2EDuration="2m5.272449314s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:25.191194209 +0000 UTC m=+145.091785914" watchObservedRunningTime="2025-11-24 06:57:25.272449314 +0000 UTC m=+145.173041019" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.312719 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" podStartSLOduration=125.312703517 podStartE2EDuration="2m5.312703517s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:25.274508365 +0000 UTC m=+145.175100070" watchObservedRunningTime="2025-11-24 06:57:25.312703517 +0000 UTC m=+145.213295222" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.325433 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7s9hl" event={"ID":"c07f2b54-9258-4f6d-bf9f-7f28e4331a84","Type":"ContainerStarted","Data":"2e2d8a94ae733ed1b026274db4a415cf4030d0017c0a022330c3527160f3f260"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.325472 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7s9hl" event={"ID":"c07f2b54-9258-4f6d-bf9f-7f28e4331a84","Type":"ContainerStarted","Data":"50b7339f6c0ea7016e19196e6cbd807ab6bd5ee8f956b51bc865d0e1c642a5ac"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.363473 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7s9hl" podStartSLOduration=125.363455676 podStartE2EDuration="2m5.363455676s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:25.361642943 +0000 UTC m=+145.262234648" watchObservedRunningTime="2025-11-24 06:57:25.363455676 +0000 UTC m=+145.264047381" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.366389 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-72wvs" event={"ID":"f1a1cfe0-beab-41d4-a478-668f5e7386f9","Type":"ContainerStarted","Data":"3858e4cd208f7a75df365ae615096b4a157cf6c04d163543ee54159afb381adc"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.367978 4809 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-72wvs container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.368022 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-72wvs" podUID="f1a1cfe0-beab-41d4-a478-668f5e7386f9" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.371490 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:25 crc kubenswrapper[4809]: E1124 06:57:25.372857 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:25.872845292 +0000 UTC m=+145.773436997 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.403662 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-56j57" event={"ID":"eca6947b-dd5e-449d-a8c5-05277c556bd1","Type":"ContainerStarted","Data":"8c9e32372254f8f62930452253b9f9f040378737fa6f5ec8ea63fe3d694a25e3"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.403701 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-56j57" event={"ID":"eca6947b-dd5e-449d-a8c5-05277c556bd1","Type":"ContainerStarted","Data":"f9d2b4a3400a2bda3a14ef6354171697dd47b38326aa591bc82fd47ef20d0695"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.422147 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8r5vw" event={"ID":"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884","Type":"ContainerStarted","Data":"a167cf5d10e05e3dc032dd5a98a000f04efc29c0ee74ed9e382c4405089d041f"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.430687 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ggzn" event={"ID":"1c002f86-a984-4feb-b110-35aebfee516e","Type":"ContainerStarted","Data":"8aede0d37d6478f0805987f77324deceb8651738a0f812f5ccc86b143d71fd47"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.440957 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-cb66b" event={"ID":"34fdb8d0-057e-40b9-a765-3e6a456ed218","Type":"ContainerStarted","Data":"63351da93f9ebca8dd78c6c34d38fc30d27b6cb0da9d0da6083a145f9207fda9"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.441057 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-cb66b" event={"ID":"34fdb8d0-057e-40b9-a765-3e6a456ed218","Type":"ContainerStarted","Data":"8ab4252238467d21c9174b99cfbb3043a9ccaec8b9b5b89aad8708e34eb23e95"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.448509 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-56j57" podStartSLOduration=125.448493982 podStartE2EDuration="2m5.448493982s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:25.447275477 +0000 UTC m=+145.347867182" watchObservedRunningTime="2025-11-24 06:57:25.448493982 +0000 UTC m=+145.349085687" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.472436 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:25 crc kubenswrapper[4809]: E1124 06:57:25.473223 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:25.973200608 +0000 UTC m=+145.873792313 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.489434 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q9xpc" event={"ID":"4ce90879-8e80-46ec-9f0b-f972764812da","Type":"ContainerStarted","Data":"6d94320fde9a61b8b36ac7154794f5e913a50cb4c1d45e71f31eabac61c0d89b"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.504898 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ggzn" podStartSLOduration=125.504875867 podStartE2EDuration="2m5.504875867s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:25.488419664 +0000 UTC m=+145.389011369" watchObservedRunningTime="2025-11-24 06:57:25.504875867 +0000 UTC m=+145.405467562" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.508319 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-54rmc" event={"ID":"96284acd-e6e2-467c-8c70-005ca153dda3","Type":"ContainerStarted","Data":"c8713bd58f199c7edc14e060e91f9adff86e0cb537e043f375adc3254ecb6323"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.526225 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-cb66b" podStartSLOduration=6.526209824 podStartE2EDuration="6.526209824s" podCreationTimestamp="2025-11-24 06:57:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:25.525138783 +0000 UTC m=+145.425730488" watchObservedRunningTime="2025-11-24 06:57:25.526209824 +0000 UTC m=+145.426801529" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.538546 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-jx6qd" event={"ID":"5fc6fa0b-133a-493e-a728-3734b122fe01","Type":"ContainerStarted","Data":"0aac84917a182ac99265dcc6127aa97d52ccb709f1740618ef0cb1285a055564"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.576450 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:25 crc kubenswrapper[4809]: E1124 06:57:25.577696 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:26.077683754 +0000 UTC m=+145.978275459 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.603611 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-8r5vw" podStartSLOduration=125.603564885 podStartE2EDuration="2m5.603564885s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:25.599465534 +0000 UTC m=+145.500057229" watchObservedRunningTime="2025-11-24 06:57:25.603564885 +0000 UTC m=+145.504156590" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.630881 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzwg5" event={"ID":"d0b8935a-d42d-4f07-876b-67f7e1ce3d89","Type":"ContainerStarted","Data":"dad317581ccebf5eac785a78549c894854fbc9fc5c60df4b11c42e115614090d"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.631188 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzwg5" event={"ID":"d0b8935a-d42d-4f07-876b-67f7e1ce3d89","Type":"ContainerStarted","Data":"4a239a1e896b0b79e76bf7868a6be04e0d935bd17dba6f401af83ce7029fd6ef"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.631273 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzwg5" event={"ID":"d0b8935a-d42d-4f07-876b-67f7e1ce3d89","Type":"ContainerStarted","Data":"52e7e87736cb1b9b58c692d7ed8c0f60014654d4cad626fb06b6e8330a5913d8"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.647875 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-54rmc" podStartSLOduration=125.647852934 podStartE2EDuration="2m5.647852934s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:25.646370321 +0000 UTC m=+145.546962026" watchObservedRunningTime="2025-11-24 06:57:25.647852934 +0000 UTC m=+145.548444639" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.651864 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-bkt2p" event={"ID":"476eceb4-f8fa-49c1-9015-5e067998af80","Type":"ContainerStarted","Data":"7c9151d6d63096bfd5e884d4181017e2d336422e70c5bf3faa413167e62b0b76"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.651991 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-bkt2p" event={"ID":"476eceb4-f8fa-49c1-9015-5e067998af80","Type":"ContainerStarted","Data":"c6c584b740382798cf36175787ddaa6614d5e0fb7719085e11c7f57f51e77a99"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.677320 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:25 crc kubenswrapper[4809]: E1124 06:57:25.677422 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:26.177406422 +0000 UTC m=+146.077998127 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.677711 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:25 crc kubenswrapper[4809]: E1124 06:57:25.679352 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:26.179341039 +0000 UTC m=+146.079932744 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.714133 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-jx6qd" podStartSLOduration=125.71411745 podStartE2EDuration="2m5.71411745s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:25.671617872 +0000 UTC m=+145.572209577" watchObservedRunningTime="2025-11-24 06:57:25.71411745 +0000 UTC m=+145.614709155" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.715715 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzwg5" podStartSLOduration=125.715705207 podStartE2EDuration="2m5.715705207s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:25.713735839 +0000 UTC m=+145.614327544" watchObservedRunningTime="2025-11-24 06:57:25.715705207 +0000 UTC m=+145.616296922" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.716034 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5w6mp" event={"ID":"e302b261-06bb-4ed0-9479-c326f26bfe43","Type":"ContainerStarted","Data":"21d53b2043a4967724b2498d822de94aa7525e3ef137b29b408e9f7614bc5d8a"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.717999 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5w6mp" event={"ID":"e302b261-06bb-4ed0-9479-c326f26bfe43","Type":"ContainerStarted","Data":"3f80ee0401dd9bf73bbcd2341127a3bcded76d6c9a5609af5baaa0872bbef293"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.718239 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5w6mp" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.722283 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" event={"ID":"47242c3c-e4f8-49f4-b508-e705396600b2","Type":"ContainerStarted","Data":"1676f36c2399cdb73b1b957c740f5061740422e713a033c0dce98c514dfeb32c"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.724105 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" event={"ID":"b08de060-f5b7-4bed-a27c-89294a8827fa","Type":"ContainerStarted","Data":"f5b1bb43ee1cfb566a2b2292ef38df7c63aeaee858c08fee970d588092fb019e"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.724187 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" event={"ID":"b08de060-f5b7-4bed-a27c-89294a8827fa","Type":"ContainerStarted","Data":"053b70f21e3f4b82b650ef00d8249129b288d2a2f87131adf31e03b86cc0ca72"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.724924 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.729288 4809 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-svxc5 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.729337 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" podUID="b08de060-f5b7-4bed-a27c-89294a8827fa" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.739647 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-wlshf" event={"ID":"053a3f86-3670-42c2-afe9-8b1f09ece542","Type":"ContainerStarted","Data":"d3b9f301f5026b0318b8671cefadaf01093d8b1309778cb98468404f2e2018d4"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.745770 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-88wkc" event={"ID":"66ef6324-33df-4e2a-92dd-36a3f5f7995d","Type":"ContainerStarted","Data":"8a664e767544624fd5de0da5407717ecd61cfd3ed82c9ba5b75dc2e1aa265ce4"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.747253 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pvrfw" event={"ID":"e846497c-2339-4f20-b688-f2483c07d54f","Type":"ContainerStarted","Data":"a0aa84e12688f33805d8ad9f1af3c3a4b68c4d9ae0624bdd18d6b465a0ae1320"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.747274 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pvrfw" event={"ID":"e846497c-2339-4f20-b688-f2483c07d54f","Type":"ContainerStarted","Data":"2fe2f8e2940300eb7aaf29c5386eb92db6e38430ae015307510c40a52dfe4c86"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.749172 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-mmdhq" event={"ID":"8dd70f5f-be62-4eda-8cb0-c197ca2bb5a4","Type":"ContainerStarted","Data":"d939b62f754b96acf89bb409cc7859c0b71cb12e40cdfd2ad7bf04a5e3538c37"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.750668 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-tk28k" event={"ID":"3cf1e8ba-0d51-4ccf-bdf1-24959b8c5c31","Type":"ContainerStarted","Data":"b3720537dea5117ebf10569fe6ba656a99c628bbb528445f59730d790e549bd3"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.759733 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xq929" event={"ID":"e3ba6e1a-6434-4867-8a19-9c9a1008b688","Type":"ContainerStarted","Data":"9d3c03a0dd1502afd7ba1172399b3a5d83bb42e58d021ab4cf94d18b48e652f6"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.783045 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:25 crc kubenswrapper[4809]: E1124 06:57:25.783586 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:26.283570599 +0000 UTC m=+146.184162304 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.783686 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:25 crc kubenswrapper[4809]: E1124 06:57:25.785620 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:26.285612109 +0000 UTC m=+146.186203814 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.788531 4809 generic.go:334] "Generic (PLEG): container finished" podID="c3b64526-f2a8-4141-b04c-65b65be09ccc" containerID="f14f45360c70c4fffa0313b6bf1a19ca015b7c7847dda26d4a56bf8a9b4569c0" exitCode=0 Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.788635 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zhgn2" event={"ID":"c3b64526-f2a8-4141-b04c-65b65be09ccc","Type":"ContainerDied","Data":"f14f45360c70c4fffa0313b6bf1a19ca015b7c7847dda26d4a56bf8a9b4569c0"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.791027 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" podStartSLOduration=125.791016526 podStartE2EDuration="2m5.791016526s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:25.789649317 +0000 UTC m=+145.690241022" watchObservedRunningTime="2025-11-24 06:57:25.791016526 +0000 UTC m=+145.691608231" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.792759 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-bkt2p" podStartSLOduration=125.792752478 podStartE2EDuration="2m5.792752478s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:25.762839559 +0000 UTC m=+145.663431284" watchObservedRunningTime="2025-11-24 06:57:25.792752478 +0000 UTC m=+145.693344183" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.808445 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mqs58" event={"ID":"7bf55845-293a-497e-87f1-eb5c36670105","Type":"ContainerStarted","Data":"2fddea131330611c0e4956bc0a1607b92d3c8f81ad107b1d00ce3506ec68dc89"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.808488 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mqs58" event={"ID":"7bf55845-293a-497e-87f1-eb5c36670105","Type":"ContainerStarted","Data":"e94466a24c46da761a412a4b9aa1ecb844d8d93e475e4d7e04a4bc997f3bf65d"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.830177 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-mmdhq" podStartSLOduration=125.830158666 podStartE2EDuration="2m5.830158666s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:25.8296298 +0000 UTC m=+145.730221505" watchObservedRunningTime="2025-11-24 06:57:25.830158666 +0000 UTC m=+145.730750371" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.870465 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-pvrfw" podStartSLOduration=125.870452269 podStartE2EDuration="2m5.870452269s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:25.868018547 +0000 UTC m=+145.768610252" watchObservedRunningTime="2025-11-24 06:57:25.870452269 +0000 UTC m=+145.771043974" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.877728 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prdvb" event={"ID":"cd9f4c00-90bd-4782-8ceb-eaab6a9799c3","Type":"ContainerStarted","Data":"ae0b0ad4bfdd680f70a5ab935b9d1cc473f3ccd2d3e3ec969d22480a07778121"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.887428 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:25 crc kubenswrapper[4809]: E1124 06:57:25.887640 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:26.387623003 +0000 UTC m=+146.288214708 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.887765 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:25 crc kubenswrapper[4809]: E1124 06:57:25.888911 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:26.38890225 +0000 UTC m=+146.289493955 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.909298 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr" event={"ID":"50767fcc-443e-40ce-84df-2182a03691f1","Type":"ContainerStarted","Data":"70cc0244033c7a813346de20a76010e8dd0081c0d82142aade810f2e84cad4f6"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.909337 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr" event={"ID":"50767fcc-443e-40ce-84df-2182a03691f1","Type":"ContainerStarted","Data":"a8024007574bd38091f8bfb0bd0b6668864704f3277fcbf2e4bfbdee2ad86aeb"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.910206 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.910531 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5w6mp" podStartSLOduration=125.910520185 podStartE2EDuration="2m5.910520185s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:25.909020711 +0000 UTC m=+145.809612416" watchObservedRunningTime="2025-11-24 06:57:25.910520185 +0000 UTC m=+145.811111890" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.915385 4809 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-r5mlr container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" start-of-body= Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.915428 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr" podUID="50767fcc-443e-40ce-84df-2182a03691f1" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.922279 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-z2dgw" event={"ID":"c1a058e6-56e0-4020-a9aa-e6537e1ccc39","Type":"ContainerStarted","Data":"bd0f5798e6e8d16e1752ae5fabd2fa7e511f8da6cc050da135aff738202e588d"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.922325 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-z2dgw" event={"ID":"c1a058e6-56e0-4020-a9aa-e6537e1ccc39","Type":"ContainerStarted","Data":"8b7e8fcdc6e38548c37a336159442815b1f369e01e2872de420ec117d5efb66e"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.972306 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pmsgp" event={"ID":"8b246b95-be2e-4c90-8cf0-f92c612bfbc1","Type":"ContainerStarted","Data":"04420973bb731c92e0ac0ccbc9fdafdfdc9a095839773b475ad80dbb41f3310c"} Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.972344 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pmsgp" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.973926 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-lkdj2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.974002 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lkdj2" podUID="b6237cc7-ba93-45a6-b560-d7916abf71f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.977143 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xq929" podStartSLOduration=125.97712776 podStartE2EDuration="2m5.97712776s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:25.976394809 +0000 UTC m=+145.876986514" watchObservedRunningTime="2025-11-24 06:57:25.97712776 +0000 UTC m=+145.877719465" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.984499 4809 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-pmsgp container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.984547 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pmsgp" podUID="8b246b95-be2e-4c90-8cf0-f92c612bfbc1" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.984670 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" Nov 24 06:57:25 crc kubenswrapper[4809]: I1124 06:57:25.988905 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:25 crc kubenswrapper[4809]: E1124 06:57:25.989305 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:26.489283527 +0000 UTC m=+146.389875232 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.006087 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr" podStartSLOduration=126.00606819 podStartE2EDuration="2m6.00606819s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:26.005276697 +0000 UTC m=+145.905868402" watchObservedRunningTime="2025-11-24 06:57:26.00606819 +0000 UTC m=+145.906659895" Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.091001 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:26 crc kubenswrapper[4809]: E1124 06:57:26.096492 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:26.596476144 +0000 UTC m=+146.497067929 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.112222 4809 patch_prober.go:28] interesting pod/router-default-5444994796-r7k56 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:57:26 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Nov 24 06:57:26 crc kubenswrapper[4809]: [+]process-running ok Nov 24 06:57:26 crc kubenswrapper[4809]: healthz check failed Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.112279 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r7k56" podUID="9eab3672-3dd0-4af1-b96d-e8c5e99bb840" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.123352 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-z2dgw" podStartSLOduration=126.123337672 podStartE2EDuration="2m6.123337672s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:26.12055528 +0000 UTC m=+146.021146995" watchObservedRunningTime="2025-11-24 06:57:26.123337672 +0000 UTC m=+146.023929377" Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.165612 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mqs58" podStartSLOduration=126.165595263 podStartE2EDuration="2m6.165595263s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:26.163814111 +0000 UTC m=+146.064405816" watchObservedRunningTime="2025-11-24 06:57:26.165595263 +0000 UTC m=+146.066186968" Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.195604 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:26 crc kubenswrapper[4809]: E1124 06:57:26.196005 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:26.695988795 +0000 UTC m=+146.596580500 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.286011 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pmsgp" podStartSLOduration=126.285994727 podStartE2EDuration="2m6.285994727s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:26.238291537 +0000 UTC m=+146.138883242" watchObservedRunningTime="2025-11-24 06:57:26.285994727 +0000 UTC m=+146.186586432" Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.287174 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prdvb" podStartSLOduration=126.287169382 podStartE2EDuration="2m6.287169382s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:26.285252065 +0000 UTC m=+146.185843770" watchObservedRunningTime="2025-11-24 06:57:26.287169382 +0000 UTC m=+146.187761087" Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.296681 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:26 crc kubenswrapper[4809]: E1124 06:57:26.296991 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:26.796980769 +0000 UTC m=+146.697572474 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.397307 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:26 crc kubenswrapper[4809]: E1124 06:57:26.397465 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:26.897444789 +0000 UTC m=+146.798036494 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.397818 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:26 crc kubenswrapper[4809]: E1124 06:57:26.398121 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:26.898110578 +0000 UTC m=+146.798702283 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.498791 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:26 crc kubenswrapper[4809]: E1124 06:57:26.498919 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:26.998895847 +0000 UTC m=+146.899487552 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.499248 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:26 crc kubenswrapper[4809]: E1124 06:57:26.499548 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:26.999537655 +0000 UTC m=+146.900129370 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.600086 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:26 crc kubenswrapper[4809]: E1124 06:57:26.600373 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:27.100334474 +0000 UTC m=+147.000926179 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.600565 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:26 crc kubenswrapper[4809]: E1124 06:57:26.600857 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:27.10084593 +0000 UTC m=+147.001437635 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.682100 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.701712 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:26 crc kubenswrapper[4809]: E1124 06:57:26.702151 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:27.202130363 +0000 UTC m=+147.102722068 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.803071 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:26 crc kubenswrapper[4809]: E1124 06:57:26.803435 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:27.303419976 +0000 UTC m=+147.204011671 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.904512 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:26 crc kubenswrapper[4809]: E1124 06:57:26.904684 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:27.404660498 +0000 UTC m=+147.305252203 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.904733 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:26 crc kubenswrapper[4809]: E1124 06:57:26.905111 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:27.40509616 +0000 UTC m=+147.305687925 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.978360 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-jx6qd" event={"ID":"5fc6fa0b-133a-493e-a728-3734b122fe01","Type":"ContainerStarted","Data":"ad9144f963afe7c105b8758712731c70852baad30382367f1a14e16ca5e23cba"} Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.980813 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-wlshf" event={"ID":"053a3f86-3670-42c2-afe9-8b1f09ece542","Type":"ContainerStarted","Data":"1146226bf9fc3418c5fdb5c658a29920d3fb082f6024626a0060b29fec7c0433"} Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.980849 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-wlshf" event={"ID":"053a3f86-3670-42c2-afe9-8b1f09ece542","Type":"ContainerStarted","Data":"6ba3121fa543578548b60eacc72984ef0571da5708861e7c3bf6df64659a62b3"} Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.981419 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-wlshf" Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.983125 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pvrfw" event={"ID":"e846497c-2339-4f20-b688-f2483c07d54f","Type":"ContainerStarted","Data":"d5197e3f8ac73a4d7cede1c722a31d520ba1528dcaab6a10b32b54395cee6ca5"} Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.984921 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5w6mp" event={"ID":"e302b261-06bb-4ed0-9479-c326f26bfe43","Type":"ContainerStarted","Data":"ba6c10072f23bde07ae5c2cd8267c53238e6339061ce6d70f95bf1ac48b770ca"} Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.987308 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zhgn2" event={"ID":"c3b64526-f2a8-4141-b04c-65b65be09ccc","Type":"ContainerStarted","Data":"3d233295679ac9de854a5ea1c15baf2ae598bdd8fdd3b754f6f9dfb830aa20e8"} Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.987450 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zhgn2" Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.989059 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" event={"ID":"04b0f650-8f86-4d8e-9fbc-860516a0f3a7","Type":"ContainerStarted","Data":"c3a183d1f9bcd690b889da25d4c71e1f00a123e4b1ca488826e76150d642cb58"} Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.991042 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q9xpc" event={"ID":"4ce90879-8e80-46ec-9f0b-f972764812da","Type":"ContainerStarted","Data":"9ce80c031e265a39e7d7156f9acad41f0fd948ff856f81374c7c67e88c969ed6"} Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.991065 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q9xpc" event={"ID":"4ce90879-8e80-46ec-9f0b-f972764812da","Type":"ContainerStarted","Data":"f8964086952a3b7b72b5e51e2cfbd952f5107c545856bfaace41ec65de6a3de6"} Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.993432 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-88wkc" event={"ID":"66ef6324-33df-4e2a-92dd-36a3f5f7995d","Type":"ContainerStarted","Data":"a8424a9a7803a95e139f50fcc11b622d323555dc01bf062d06f03f20ed532ed8"} Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.995984 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" event={"ID":"47242c3c-e4f8-49f4-b508-e705396600b2","Type":"ContainerStarted","Data":"9a9012ef7241b150a2057af8f19d8b3bbfc9021ad48f4aeab900820a90e4a8fd"} Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.997777 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-lkdj2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Nov 24 06:57:26 crc kubenswrapper[4809]: I1124 06:57:26.997814 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lkdj2" podUID="b6237cc7-ba93-45a6-b560-d7916abf71f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.003386 4809 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-svxc5 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.003440 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" podUID="b08de060-f5b7-4bed-a27c-89294a8827fa" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.004878 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-72wvs" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.006281 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:27 crc kubenswrapper[4809]: E1124 06:57:27.006673 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:27.506657501 +0000 UTC m=+147.407249206 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.016917 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pmsgp" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.019293 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-wlshf" podStartSLOduration=8.019274202 podStartE2EDuration="8.019274202s" podCreationTimestamp="2025-11-24 06:57:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:27.016378337 +0000 UTC m=+146.916970042" watchObservedRunningTime="2025-11-24 06:57:27.019274202 +0000 UTC m=+146.919865907" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.078879 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-88wkc" podStartSLOduration=127.078859651 podStartE2EDuration="2m7.078859651s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:27.074100871 +0000 UTC m=+146.974692576" watchObservedRunningTime="2025-11-24 06:57:27.078859651 +0000 UTC m=+146.979451346" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.099045 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" podStartSLOduration=127.099029413 podStartE2EDuration="2m7.099029413s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:27.096555281 +0000 UTC m=+146.997146986" watchObservedRunningTime="2025-11-24 06:57:27.099029413 +0000 UTC m=+146.999621118" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.109037 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:27 crc kubenswrapper[4809]: E1124 06:57:27.110675 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:27.610661985 +0000 UTC m=+147.511253690 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.110762 4809 patch_prober.go:28] interesting pod/router-default-5444994796-r7k56 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:57:27 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Nov 24 06:57:27 crc kubenswrapper[4809]: [+]process-running ok Nov 24 06:57:27 crc kubenswrapper[4809]: healthz check failed Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.110797 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r7k56" podUID="9eab3672-3dd0-4af1-b96d-e8c5e99bb840" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.126069 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-q9xpc" podStartSLOduration=127.126047006 podStartE2EDuration="2m7.126047006s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:27.122379209 +0000 UTC m=+147.022970914" watchObservedRunningTime="2025-11-24 06:57:27.126047006 +0000 UTC m=+147.026638711" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.211783 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:27 crc kubenswrapper[4809]: E1124 06:57:27.212124 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:27.712108253 +0000 UTC m=+147.612699958 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.219556 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zhgn2" podStartSLOduration=127.219540021 podStartE2EDuration="2m7.219540021s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:27.193048943 +0000 UTC m=+147.093640648" watchObservedRunningTime="2025-11-24 06:57:27.219540021 +0000 UTC m=+147.120131726" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.313445 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:27 crc kubenswrapper[4809]: E1124 06:57:27.313858 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:27.813841109 +0000 UTC m=+147.714432814 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.416325 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:27 crc kubenswrapper[4809]: E1124 06:57:27.416634 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:27.916619216 +0000 UTC m=+147.817210921 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.518186 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:27 crc kubenswrapper[4809]: E1124 06:57:27.518572 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:28.018555638 +0000 UTC m=+147.919147343 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.619418 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:27 crc kubenswrapper[4809]: E1124 06:57:27.619786 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:28.119768469 +0000 UTC m=+148.020360174 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.619875 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:27 crc kubenswrapper[4809]: E1124 06:57:27.620214 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:28.120206452 +0000 UTC m=+148.020798157 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.653647 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pj46g"] Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.654576 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pj46g" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.658363 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.689392 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pj46g"] Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.721283 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:27 crc kubenswrapper[4809]: E1124 06:57:27.721430 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:28.221412994 +0000 UTC m=+148.122004699 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.721518 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk262\" (UniqueName: \"kubernetes.io/projected/1383ed81-5c2a-4416-8c67-05e4881db7be-kube-api-access-hk262\") pod \"certified-operators-pj46g\" (UID: \"1383ed81-5c2a-4416-8c67-05e4881db7be\") " pod="openshift-marketplace/certified-operators-pj46g" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.721550 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.721577 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1383ed81-5c2a-4416-8c67-05e4881db7be-catalog-content\") pod \"certified-operators-pj46g\" (UID: \"1383ed81-5c2a-4416-8c67-05e4881db7be\") " pod="openshift-marketplace/certified-operators-pj46g" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.721616 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1383ed81-5c2a-4416-8c67-05e4881db7be-utilities\") pod \"certified-operators-pj46g\" (UID: \"1383ed81-5c2a-4416-8c67-05e4881db7be\") " pod="openshift-marketplace/certified-operators-pj46g" Nov 24 06:57:27 crc kubenswrapper[4809]: E1124 06:57:27.721788 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:28.221779264 +0000 UTC m=+148.122370969 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.822151 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:27 crc kubenswrapper[4809]: E1124 06:57:27.822306 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:28.322280925 +0000 UTC m=+148.222872630 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.822353 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.822386 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1383ed81-5c2a-4416-8c67-05e4881db7be-catalog-content\") pod \"certified-operators-pj46g\" (UID: \"1383ed81-5c2a-4416-8c67-05e4881db7be\") " pod="openshift-marketplace/certified-operators-pj46g" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.822428 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1383ed81-5c2a-4416-8c67-05e4881db7be-utilities\") pod \"certified-operators-pj46g\" (UID: \"1383ed81-5c2a-4416-8c67-05e4881db7be\") " pod="openshift-marketplace/certified-operators-pj46g" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.822503 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk262\" (UniqueName: \"kubernetes.io/projected/1383ed81-5c2a-4416-8c67-05e4881db7be-kube-api-access-hk262\") pod \"certified-operators-pj46g\" (UID: \"1383ed81-5c2a-4416-8c67-05e4881db7be\") " pod="openshift-marketplace/certified-operators-pj46g" Nov 24 06:57:27 crc kubenswrapper[4809]: E1124 06:57:27.822656 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:28.322648995 +0000 UTC m=+148.223240700 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.822771 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1383ed81-5c2a-4416-8c67-05e4881db7be-catalog-content\") pod \"certified-operators-pj46g\" (UID: \"1383ed81-5c2a-4416-8c67-05e4881db7be\") " pod="openshift-marketplace/certified-operators-pj46g" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.822814 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1383ed81-5c2a-4416-8c67-05e4881db7be-utilities\") pod \"certified-operators-pj46g\" (UID: \"1383ed81-5c2a-4416-8c67-05e4881db7be\") " pod="openshift-marketplace/certified-operators-pj46g" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.841774 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-llxlp"] Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.842636 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-llxlp" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.848246 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.849001 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk262\" (UniqueName: \"kubernetes.io/projected/1383ed81-5c2a-4416-8c67-05e4881db7be-kube-api-access-hk262\") pod \"certified-operators-pj46g\" (UID: \"1383ed81-5c2a-4416-8c67-05e4881db7be\") " pod="openshift-marketplace/certified-operators-pj46g" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.858414 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-llxlp"] Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.924057 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:27 crc kubenswrapper[4809]: E1124 06:57:27.924227 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:28.424200326 +0000 UTC m=+148.324792031 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.924297 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8951296-b323-4ac3-9e01-963b60ff998c-utilities\") pod \"community-operators-llxlp\" (UID: \"f8951296-b323-4ac3-9e01-963b60ff998c\") " pod="openshift-marketplace/community-operators-llxlp" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.924328 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.924409 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8951296-b323-4ac3-9e01-963b60ff998c-catalog-content\") pod \"community-operators-llxlp\" (UID: \"f8951296-b323-4ac3-9e01-963b60ff998c\") " pod="openshift-marketplace/community-operators-llxlp" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.924446 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58z8t\" (UniqueName: \"kubernetes.io/projected/f8951296-b323-4ac3-9e01-963b60ff998c-kube-api-access-58z8t\") pod \"community-operators-llxlp\" (UID: \"f8951296-b323-4ac3-9e01-963b60ff998c\") " pod="openshift-marketplace/community-operators-llxlp" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.924486 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.924518 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:57:27 crc kubenswrapper[4809]: E1124 06:57:27.924732 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:28.424720591 +0000 UTC m=+148.325312296 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.928073 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.929404 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.966410 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pj46g" Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.999079 4809 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-r5mlr container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 06:57:27 crc kubenswrapper[4809]: I1124 06:57:27.999133 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr" podUID="50767fcc-443e-40ce-84df-2182a03691f1" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.003123 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" event={"ID":"47242c3c-e4f8-49f4-b508-e705396600b2","Type":"ContainerStarted","Data":"43a974721f0cdd99ed106e21e4e6060593cd0b74a64380894317a0d39d054957"} Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.007181 4809 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-svxc5 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.007236 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" podUID="b08de060-f5b7-4bed-a27c-89294a8827fa" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.025426 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.025541 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8951296-b323-4ac3-9e01-963b60ff998c-utilities\") pod \"community-operators-llxlp\" (UID: \"f8951296-b323-4ac3-9e01-963b60ff998c\") " pod="openshift-marketplace/community-operators-llxlp" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.025663 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.025681 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.025741 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8951296-b323-4ac3-9e01-963b60ff998c-catalog-content\") pod \"community-operators-llxlp\" (UID: \"f8951296-b323-4ac3-9e01-963b60ff998c\") " pod="openshift-marketplace/community-operators-llxlp" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.025890 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58z8t\" (UniqueName: \"kubernetes.io/projected/f8951296-b323-4ac3-9e01-963b60ff998c-kube-api-access-58z8t\") pod \"community-operators-llxlp\" (UID: \"f8951296-b323-4ac3-9e01-963b60ff998c\") " pod="openshift-marketplace/community-operators-llxlp" Nov 24 06:57:28 crc kubenswrapper[4809]: E1124 06:57:28.027455 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:28.527433417 +0000 UTC m=+148.428025122 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.028158 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8951296-b323-4ac3-9e01-963b60ff998c-utilities\") pod \"community-operators-llxlp\" (UID: \"f8951296-b323-4ac3-9e01-963b60ff998c\") " pod="openshift-marketplace/community-operators-llxlp" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.028709 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8951296-b323-4ac3-9e01-963b60ff998c-catalog-content\") pod \"community-operators-llxlp\" (UID: \"f8951296-b323-4ac3-9e01-963b60ff998c\") " pod="openshift-marketplace/community-operators-llxlp" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.034301 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.038554 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.041935 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bnt6m"] Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.042821 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bnt6m" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.053664 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58z8t\" (UniqueName: \"kubernetes.io/projected/f8951296-b323-4ac3-9e01-963b60ff998c-kube-api-access-58z8t\") pod \"community-operators-llxlp\" (UID: \"f8951296-b323-4ac3-9e01-963b60ff998c\") " pod="openshift-marketplace/community-operators-llxlp" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.055940 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bnt6m"] Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.113097 4809 patch_prober.go:28] interesting pod/router-default-5444994796-r7k56 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:57:28 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Nov 24 06:57:28 crc kubenswrapper[4809]: [+]process-running ok Nov 24 06:57:28 crc kubenswrapper[4809]: healthz check failed Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.113150 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r7k56" podUID="9eab3672-3dd0-4af1-b96d-e8c5e99bb840" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.126844 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:28 crc kubenswrapper[4809]: E1124 06:57:28.127170 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:28.627156053 +0000 UTC m=+148.527747758 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.187236 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-llxlp" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.219278 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.227519 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.227722 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qglxv\" (UniqueName: \"kubernetes.io/projected/fb64a807-42bd-4297-aa31-52a142255d89-kube-api-access-qglxv\") pod \"certified-operators-bnt6m\" (UID: \"fb64a807-42bd-4297-aa31-52a142255d89\") " pod="openshift-marketplace/certified-operators-bnt6m" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.227779 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb64a807-42bd-4297-aa31-52a142255d89-catalog-content\") pod \"certified-operators-bnt6m\" (UID: \"fb64a807-42bd-4297-aa31-52a142255d89\") " pod="openshift-marketplace/certified-operators-bnt6m" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.227813 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb64a807-42bd-4297-aa31-52a142255d89-utilities\") pod \"certified-operators-bnt6m\" (UID: \"fb64a807-42bd-4297-aa31-52a142255d89\") " pod="openshift-marketplace/certified-operators-bnt6m" Nov 24 06:57:28 crc kubenswrapper[4809]: E1124 06:57:28.227991 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:28.727954242 +0000 UTC m=+148.628545947 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.249676 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9ttsc"] Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.250755 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9ttsc" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.263443 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9ttsc"] Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.313226 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.326024 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.332030 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb64a807-42bd-4297-aa31-52a142255d89-catalog-content\") pod \"certified-operators-bnt6m\" (UID: \"fb64a807-42bd-4297-aa31-52a142255d89\") " pod="openshift-marketplace/certified-operators-bnt6m" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.332078 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c69d61e-4b0d-4388-aff9-9571d2a9c16c-catalog-content\") pod \"community-operators-9ttsc\" (UID: \"8c69d61e-4b0d-4388-aff9-9571d2a9c16c\") " pod="openshift-marketplace/community-operators-9ttsc" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.332112 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb64a807-42bd-4297-aa31-52a142255d89-utilities\") pod \"certified-operators-bnt6m\" (UID: \"fb64a807-42bd-4297-aa31-52a142255d89\") " pod="openshift-marketplace/certified-operators-bnt6m" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.332222 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c69d61e-4b0d-4388-aff9-9571d2a9c16c-utilities\") pod \"community-operators-9ttsc\" (UID: \"8c69d61e-4b0d-4388-aff9-9571d2a9c16c\") " pod="openshift-marketplace/community-operators-9ttsc" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.332321 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vm75s\" (UniqueName: \"kubernetes.io/projected/8c69d61e-4b0d-4388-aff9-9571d2a9c16c-kube-api-access-vm75s\") pod \"community-operators-9ttsc\" (UID: \"8c69d61e-4b0d-4388-aff9-9571d2a9c16c\") " pod="openshift-marketplace/community-operators-9ttsc" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.332411 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.332475 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qglxv\" (UniqueName: \"kubernetes.io/projected/fb64a807-42bd-4297-aa31-52a142255d89-kube-api-access-qglxv\") pod \"certified-operators-bnt6m\" (UID: \"fb64a807-42bd-4297-aa31-52a142255d89\") " pod="openshift-marketplace/certified-operators-bnt6m" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.332575 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb64a807-42bd-4297-aa31-52a142255d89-catalog-content\") pod \"certified-operators-bnt6m\" (UID: \"fb64a807-42bd-4297-aa31-52a142255d89\") " pod="openshift-marketplace/certified-operators-bnt6m" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.332649 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb64a807-42bd-4297-aa31-52a142255d89-utilities\") pod \"certified-operators-bnt6m\" (UID: \"fb64a807-42bd-4297-aa31-52a142255d89\") " pod="openshift-marketplace/certified-operators-bnt6m" Nov 24 06:57:28 crc kubenswrapper[4809]: E1124 06:57:28.332870 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:28.832857512 +0000 UTC m=+148.733449217 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.357910 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qglxv\" (UniqueName: \"kubernetes.io/projected/fb64a807-42bd-4297-aa31-52a142255d89-kube-api-access-qglxv\") pod \"certified-operators-bnt6m\" (UID: \"fb64a807-42bd-4297-aa31-52a142255d89\") " pod="openshift-marketplace/certified-operators-bnt6m" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.364195 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bnt6m" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.433466 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.433661 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vm75s\" (UniqueName: \"kubernetes.io/projected/8c69d61e-4b0d-4388-aff9-9571d2a9c16c-kube-api-access-vm75s\") pod \"community-operators-9ttsc\" (UID: \"8c69d61e-4b0d-4388-aff9-9571d2a9c16c\") " pod="openshift-marketplace/community-operators-9ttsc" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.433738 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c69d61e-4b0d-4388-aff9-9571d2a9c16c-catalog-content\") pod \"community-operators-9ttsc\" (UID: \"8c69d61e-4b0d-4388-aff9-9571d2a9c16c\") " pod="openshift-marketplace/community-operators-9ttsc" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.433952 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c69d61e-4b0d-4388-aff9-9571d2a9c16c-utilities\") pod \"community-operators-9ttsc\" (UID: \"8c69d61e-4b0d-4388-aff9-9571d2a9c16c\") " pod="openshift-marketplace/community-operators-9ttsc" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.434370 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c69d61e-4b0d-4388-aff9-9571d2a9c16c-utilities\") pod \"community-operators-9ttsc\" (UID: \"8c69d61e-4b0d-4388-aff9-9571d2a9c16c\") " pod="openshift-marketplace/community-operators-9ttsc" Nov 24 06:57:28 crc kubenswrapper[4809]: E1124 06:57:28.434436 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:28.934421553 +0000 UTC m=+148.835013258 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.434878 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c69d61e-4b0d-4388-aff9-9571d2a9c16c-catalog-content\") pod \"community-operators-9ttsc\" (UID: \"8c69d61e-4b0d-4388-aff9-9571d2a9c16c\") " pod="openshift-marketplace/community-operators-9ttsc" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.473351 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vm75s\" (UniqueName: \"kubernetes.io/projected/8c69d61e-4b0d-4388-aff9-9571d2a9c16c-kube-api-access-vm75s\") pod \"community-operators-9ttsc\" (UID: \"8c69d61e-4b0d-4388-aff9-9571d2a9c16c\") " pod="openshift-marketplace/community-operators-9ttsc" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.543729 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:28 crc kubenswrapper[4809]: E1124 06:57:28.544306 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:29.044295129 +0000 UTC m=+148.944886834 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.547556 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r5mlr" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.568122 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pj46g"] Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.587250 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9ttsc" Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.645161 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:28 crc kubenswrapper[4809]: E1124 06:57:28.645442 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:29.145426887 +0000 UTC m=+149.046018592 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.646154 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-llxlp"] Nov 24 06:57:28 crc kubenswrapper[4809]: W1124 06:57:28.672445 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1383ed81_5c2a_4416_8c67_05e4881db7be.slice/crio-66c7cd27d613d48d506293c04b3f19db0acfa6edb3398e30423b3945cb41f9a3 WatchSource:0}: Error finding container 66c7cd27d613d48d506293c04b3f19db0acfa6edb3398e30423b3945cb41f9a3: Status 404 returned error can't find the container with id 66c7cd27d613d48d506293c04b3f19db0acfa6edb3398e30423b3945cb41f9a3 Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.747139 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:28 crc kubenswrapper[4809]: E1124 06:57:28.747461 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:29.247449422 +0000 UTC m=+149.148041127 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.848212 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:28 crc kubenswrapper[4809]: E1124 06:57:28.848929 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:29.348913721 +0000 UTC m=+149.249505416 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.934902 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bnt6m"] Nov 24 06:57:28 crc kubenswrapper[4809]: I1124 06:57:28.951247 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:28 crc kubenswrapper[4809]: E1124 06:57:28.951599 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:29.451583915 +0000 UTC m=+149.352175620 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.008082 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9ttsc"] Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.011626 4809 generic.go:334] "Generic (PLEG): container finished" podID="f8951296-b323-4ac3-9e01-963b60ff998c" containerID="064add890a06df5449f17b0d1227089f5f034abca0fc1bdae562272bd597ba2d" exitCode=0 Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.011682 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llxlp" event={"ID":"f8951296-b323-4ac3-9e01-963b60ff998c","Type":"ContainerDied","Data":"064add890a06df5449f17b0d1227089f5f034abca0fc1bdae562272bd597ba2d"} Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.011706 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llxlp" event={"ID":"f8951296-b323-4ac3-9e01-963b60ff998c","Type":"ContainerStarted","Data":"10efd5615b838c8f298a6c3da72c970496b9cb66d0fcbfc009a579b7e4164ff0"} Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.014003 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.017708 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" event={"ID":"47242c3c-e4f8-49f4-b508-e705396600b2","Type":"ContainerStarted","Data":"4b4233e47126eea4bd7affdc063b3596ac88beb49ec4b2ea63b42bd51d621eb1"} Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.017752 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" event={"ID":"47242c3c-e4f8-49f4-b508-e705396600b2","Type":"ContainerStarted","Data":"bbb8e4a271cc57f0f48141071828badd58817b331858ab1d640daed8e8fce9c6"} Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.019117 4809 generic.go:334] "Generic (PLEG): container finished" podID="1383ed81-5c2a-4416-8c67-05e4881db7be" containerID="d41fd86b045fd2a17480ad649f2ddc752efbf24b64eb1325608f4f1c9bc9b5ae" exitCode=0 Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.019162 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pj46g" event={"ID":"1383ed81-5c2a-4416-8c67-05e4881db7be","Type":"ContainerDied","Data":"d41fd86b045fd2a17480ad649f2ddc752efbf24b64eb1325608f4f1c9bc9b5ae"} Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.019177 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pj46g" event={"ID":"1383ed81-5c2a-4416-8c67-05e4881db7be","Type":"ContainerStarted","Data":"66c7cd27d613d48d506293c04b3f19db0acfa6edb3398e30423b3945cb41f9a3"} Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.023538 4809 generic.go:334] "Generic (PLEG): container finished" podID="476eceb4-f8fa-49c1-9015-5e067998af80" containerID="7c9151d6d63096bfd5e884d4181017e2d336422e70c5bf3faa413167e62b0b76" exitCode=0 Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.023703 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-bkt2p" event={"ID":"476eceb4-f8fa-49c1-9015-5e067998af80","Type":"ContainerDied","Data":"7c9151d6d63096bfd5e884d4181017e2d336422e70c5bf3faa413167e62b0b76"} Nov 24 06:57:29 crc kubenswrapper[4809]: W1124 06:57:29.027336 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb64a807_42bd_4297_aa31_52a142255d89.slice/crio-b6b42b0a5363a81f8bb60d7c0456d5df702d747cff9b7e0ae92c835cb677debd WatchSource:0}: Error finding container b6b42b0a5363a81f8bb60d7c0456d5df702d747cff9b7e0ae92c835cb677debd: Status 404 returned error can't find the container with id b6b42b0a5363a81f8bb60d7c0456d5df702d747cff9b7e0ae92c835cb677debd Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.052163 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:29 crc kubenswrapper[4809]: E1124 06:57:29.052304 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:29.552283651 +0000 UTC m=+149.452875356 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.052394 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:29 crc kubenswrapper[4809]: E1124 06:57:29.052833 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:29.552817307 +0000 UTC m=+149.453409012 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.061151 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-gzp8q" podStartSLOduration=10.061130591 podStartE2EDuration="10.061130591s" podCreationTimestamp="2025-11-24 06:57:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:29.059847372 +0000 UTC m=+148.960439077" watchObservedRunningTime="2025-11-24 06:57:29.061130591 +0000 UTC m=+148.961722296" Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.113467 4809 patch_prober.go:28] interesting pod/router-default-5444994796-r7k56 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:57:29 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Nov 24 06:57:29 crc kubenswrapper[4809]: [+]process-running ok Nov 24 06:57:29 crc kubenswrapper[4809]: healthz check failed Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.113528 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r7k56" podUID="9eab3672-3dd0-4af1-b96d-e8c5e99bb840" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.153207 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:29 crc kubenswrapper[4809]: E1124 06:57:29.153455 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:29.6534419 +0000 UTC m=+149.554033605 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.254211 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:29 crc kubenswrapper[4809]: E1124 06:57:29.254567 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:29.754551758 +0000 UTC m=+149.655143463 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.272132 4809 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.355922 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:29 crc kubenswrapper[4809]: E1124 06:57:29.356261 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:29.856242803 +0000 UTC m=+149.756834518 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.457206 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:29 crc kubenswrapper[4809]: E1124 06:57:29.457726 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:29.957705191 +0000 UTC m=+149.858296936 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.557876 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:29 crc kubenswrapper[4809]: E1124 06:57:29.558236 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:30.058218292 +0000 UTC m=+149.958810007 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.633551 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9r2kr"] Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.635197 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9r2kr" Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.638006 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.661226 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9r2kr"] Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.661433 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c707d1d-d2c2-42db-82b9-afe8fde54914-catalog-content\") pod \"redhat-marketplace-9r2kr\" (UID: \"9c707d1d-d2c2-42db-82b9-afe8fde54914\") " pod="openshift-marketplace/redhat-marketplace-9r2kr" Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.661507 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c707d1d-d2c2-42db-82b9-afe8fde54914-utilities\") pod \"redhat-marketplace-9r2kr\" (UID: \"9c707d1d-d2c2-42db-82b9-afe8fde54914\") " pod="openshift-marketplace/redhat-marketplace-9r2kr" Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.661571 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.661651 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drtpr\" (UniqueName: \"kubernetes.io/projected/9c707d1d-d2c2-42db-82b9-afe8fde54914-kube-api-access-drtpr\") pod \"redhat-marketplace-9r2kr\" (UID: \"9c707d1d-d2c2-42db-82b9-afe8fde54914\") " pod="openshift-marketplace/redhat-marketplace-9r2kr" Nov 24 06:57:29 crc kubenswrapper[4809]: E1124 06:57:29.661939 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:30.161927876 +0000 UTC m=+150.062519581 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.763343 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:29 crc kubenswrapper[4809]: E1124 06:57:29.763554 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:30.263521839 +0000 UTC m=+150.164113554 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.763733 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drtpr\" (UniqueName: \"kubernetes.io/projected/9c707d1d-d2c2-42db-82b9-afe8fde54914-kube-api-access-drtpr\") pod \"redhat-marketplace-9r2kr\" (UID: \"9c707d1d-d2c2-42db-82b9-afe8fde54914\") " pod="openshift-marketplace/redhat-marketplace-9r2kr" Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.763855 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c707d1d-d2c2-42db-82b9-afe8fde54914-catalog-content\") pod \"redhat-marketplace-9r2kr\" (UID: \"9c707d1d-d2c2-42db-82b9-afe8fde54914\") " pod="openshift-marketplace/redhat-marketplace-9r2kr" Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.763952 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c707d1d-d2c2-42db-82b9-afe8fde54914-utilities\") pod \"redhat-marketplace-9r2kr\" (UID: \"9c707d1d-d2c2-42db-82b9-afe8fde54914\") " pod="openshift-marketplace/redhat-marketplace-9r2kr" Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.764337 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c707d1d-d2c2-42db-82b9-afe8fde54914-catalog-content\") pod \"redhat-marketplace-9r2kr\" (UID: \"9c707d1d-d2c2-42db-82b9-afe8fde54914\") " pod="openshift-marketplace/redhat-marketplace-9r2kr" Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.764491 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c707d1d-d2c2-42db-82b9-afe8fde54914-utilities\") pod \"redhat-marketplace-9r2kr\" (UID: \"9c707d1d-d2c2-42db-82b9-afe8fde54914\") " pod="openshift-marketplace/redhat-marketplace-9r2kr" Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.788145 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drtpr\" (UniqueName: \"kubernetes.io/projected/9c707d1d-d2c2-42db-82b9-afe8fde54914-kube-api-access-drtpr\") pod \"redhat-marketplace-9r2kr\" (UID: \"9c707d1d-d2c2-42db-82b9-afe8fde54914\") " pod="openshift-marketplace/redhat-marketplace-9r2kr" Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.865615 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:29 crc kubenswrapper[4809]: E1124 06:57:29.865906 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:30.365894794 +0000 UTC m=+150.266486499 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.952187 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9r2kr" Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.966567 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:29 crc kubenswrapper[4809]: E1124 06:57:29.966743 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:30.466716153 +0000 UTC m=+150.367307858 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:29 crc kubenswrapper[4809]: I1124 06:57:29.966846 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:29 crc kubenswrapper[4809]: E1124 06:57:29.967156 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:57:30.467144806 +0000 UTC m=+150.367736511 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-59cgv" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.034416 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7rrb4"] Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.035589 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7rrb4" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.037489 4809 generic.go:334] "Generic (PLEG): container finished" podID="fb64a807-42bd-4297-aa31-52a142255d89" containerID="2be25d53b63d196c5855c6d2768759424b8b21f2b5c00ec3ed4aa14a2da35c4a" exitCode=0 Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.038369 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bnt6m" event={"ID":"fb64a807-42bd-4297-aa31-52a142255d89","Type":"ContainerDied","Data":"2be25d53b63d196c5855c6d2768759424b8b21f2b5c00ec3ed4aa14a2da35c4a"} Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.038417 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bnt6m" event={"ID":"fb64a807-42bd-4297-aa31-52a142255d89","Type":"ContainerStarted","Data":"b6b42b0a5363a81f8bb60d7c0456d5df702d747cff9b7e0ae92c835cb677debd"} Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.041681 4809 generic.go:334] "Generic (PLEG): container finished" podID="8c69d61e-4b0d-4388-aff9-9571d2a9c16c" containerID="d00b465a0808d4b37f877b29f106cf72c4d30f0410672f93e34a2c3d6c1c2912" exitCode=0 Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.041744 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9ttsc" event={"ID":"8c69d61e-4b0d-4388-aff9-9571d2a9c16c","Type":"ContainerDied","Data":"d00b465a0808d4b37f877b29f106cf72c4d30f0410672f93e34a2c3d6c1c2912"} Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.041771 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9ttsc" event={"ID":"8c69d61e-4b0d-4388-aff9-9571d2a9c16c","Type":"ContainerStarted","Data":"31f65ad8a0d43722450c775e01cc7a8f016e1f8c04bc5ab6b856b179b7d7da1c"} Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.044158 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7rrb4"] Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.045798 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"468c87dd493cc1b8ef28d4719e585712831db3bf95c16e07b99968cf17f259f6"} Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.045821 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7119d04ba9ab1ac71fb9ef54b71e335750177b02e35379cf3f9e2a1c15ed7543"} Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.047529 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"07c372b2112af14f136c0e36a287bd63a4dc2ca521f921b094e757206d223592"} Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.047563 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"0351afd94ce21b904bd942b89c721d1502d7de6a6a7cd4c629cf00205f335a78"} Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.049270 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"72ac04bc38f868df4a5e0eaf98ba0e51256ef166019f8e389fb54c923ed9cd87"} Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.049354 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"df9f57cb6536554eebd55e97a1d33871258c8a840d460e1105cc6b5d016a9b63"} Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.049952 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.077415 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.077715 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcbrr\" (UniqueName: \"kubernetes.io/projected/c385307d-febf-4fe4-bdf6-51224328ea10-kube-api-access-kcbrr\") pod \"redhat-marketplace-7rrb4\" (UID: \"c385307d-febf-4fe4-bdf6-51224328ea10\") " pod="openshift-marketplace/redhat-marketplace-7rrb4" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.077741 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c385307d-febf-4fe4-bdf6-51224328ea10-utilities\") pod \"redhat-marketplace-7rrb4\" (UID: \"c385307d-febf-4fe4-bdf6-51224328ea10\") " pod="openshift-marketplace/redhat-marketplace-7rrb4" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.077789 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c385307d-febf-4fe4-bdf6-51224328ea10-catalog-content\") pod \"redhat-marketplace-7rrb4\" (UID: \"c385307d-febf-4fe4-bdf6-51224328ea10\") " pod="openshift-marketplace/redhat-marketplace-7rrb4" Nov 24 06:57:30 crc kubenswrapper[4809]: E1124 06:57:30.078368 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:57:30.578349741 +0000 UTC m=+150.478941446 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.082249 4809 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-24T06:57:29.272165685Z","Handler":null,"Name":""} Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.098434 4809 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.098482 4809 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.114322 4809 patch_prober.go:28] interesting pod/router-default-5444994796-r7k56 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:57:30 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Nov 24 06:57:30 crc kubenswrapper[4809]: [+]process-running ok Nov 24 06:57:30 crc kubenswrapper[4809]: healthz check failed Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.114372 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r7k56" podUID="9eab3672-3dd0-4af1-b96d-e8c5e99bb840" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.180468 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcbrr\" (UniqueName: \"kubernetes.io/projected/c385307d-febf-4fe4-bdf6-51224328ea10-kube-api-access-kcbrr\") pod \"redhat-marketplace-7rrb4\" (UID: \"c385307d-febf-4fe4-bdf6-51224328ea10\") " pod="openshift-marketplace/redhat-marketplace-7rrb4" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.180497 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c385307d-febf-4fe4-bdf6-51224328ea10-utilities\") pod \"redhat-marketplace-7rrb4\" (UID: \"c385307d-febf-4fe4-bdf6-51224328ea10\") " pod="openshift-marketplace/redhat-marketplace-7rrb4" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.180531 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c385307d-febf-4fe4-bdf6-51224328ea10-catalog-content\") pod \"redhat-marketplace-7rrb4\" (UID: \"c385307d-febf-4fe4-bdf6-51224328ea10\") " pod="openshift-marketplace/redhat-marketplace-7rrb4" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.180565 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.181692 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c385307d-febf-4fe4-bdf6-51224328ea10-utilities\") pod \"redhat-marketplace-7rrb4\" (UID: \"c385307d-febf-4fe4-bdf6-51224328ea10\") " pod="openshift-marketplace/redhat-marketplace-7rrb4" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.181902 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c385307d-febf-4fe4-bdf6-51224328ea10-catalog-content\") pod \"redhat-marketplace-7rrb4\" (UID: \"c385307d-febf-4fe4-bdf6-51224328ea10\") " pod="openshift-marketplace/redhat-marketplace-7rrb4" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.205751 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.205792 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.244844 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcbrr\" (UniqueName: \"kubernetes.io/projected/c385307d-febf-4fe4-bdf6-51224328ea10-kube-api-access-kcbrr\") pod \"redhat-marketplace-7rrb4\" (UID: \"c385307d-febf-4fe4-bdf6-51224328ea10\") " pod="openshift-marketplace/redhat-marketplace-7rrb4" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.314530 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-59cgv\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.336844 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.358393 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-bkt2p" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.382476 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/476eceb4-f8fa-49c1-9015-5e067998af80-secret-volume\") pod \"476eceb4-f8fa-49c1-9015-5e067998af80\" (UID: \"476eceb4-f8fa-49c1-9015-5e067998af80\") " Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.382522 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szl84\" (UniqueName: \"kubernetes.io/projected/476eceb4-f8fa-49c1-9015-5e067998af80-kube-api-access-szl84\") pod \"476eceb4-f8fa-49c1-9015-5e067998af80\" (UID: \"476eceb4-f8fa-49c1-9015-5e067998af80\") " Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.382662 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/476eceb4-f8fa-49c1-9015-5e067998af80-config-volume\") pod \"476eceb4-f8fa-49c1-9015-5e067998af80\" (UID: \"476eceb4-f8fa-49c1-9015-5e067998af80\") " Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.382769 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.383300 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/476eceb4-f8fa-49c1-9015-5e067998af80-config-volume" (OuterVolumeSpecName: "config-volume") pod "476eceb4-f8fa-49c1-9015-5e067998af80" (UID: "476eceb4-f8fa-49c1-9015-5e067998af80"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.388941 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/476eceb4-f8fa-49c1-9015-5e067998af80-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "476eceb4-f8fa-49c1-9015-5e067998af80" (UID: "476eceb4-f8fa-49c1-9015-5e067998af80"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.389741 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/476eceb4-f8fa-49c1-9015-5e067998af80-kube-api-access-szl84" (OuterVolumeSpecName: "kube-api-access-szl84") pod "476eceb4-f8fa-49c1-9015-5e067998af80" (UID: "476eceb4-f8fa-49c1-9015-5e067998af80"). InnerVolumeSpecName "kube-api-access-szl84". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.390710 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.415180 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7rrb4" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.484842 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/476eceb4-f8fa-49c1-9015-5e067998af80-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.484868 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/476eceb4-f8fa-49c1-9015-5e067998af80-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.484880 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szl84\" (UniqueName: \"kubernetes.io/projected/476eceb4-f8fa-49c1-9015-5e067998af80-kube-api-access-szl84\") on node \"crc\" DevicePath \"\"" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.537434 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 06:57:30 crc kubenswrapper[4809]: E1124 06:57:30.537622 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="476eceb4-f8fa-49c1-9015-5e067998af80" containerName="collect-profiles" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.537632 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="476eceb4-f8fa-49c1-9015-5e067998af80" containerName="collect-profiles" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.537743 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="476eceb4-f8fa-49c1-9015-5e067998af80" containerName="collect-profiles" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.538097 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.540486 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.540560 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.547890 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.554155 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-59cgv"] Nov 24 06:57:30 crc kubenswrapper[4809]: W1124 06:57:30.583213 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4af1418f_a56e_4f26_a800_3b402b7da053.slice/crio-5921c15b8acc0634b00fa4668ce6b4dee98621c0f70d803226c323753698c5ce WatchSource:0}: Error finding container 5921c15b8acc0634b00fa4668ce6b4dee98621c0f70d803226c323753698c5ce: Status 404 returned error can't find the container with id 5921c15b8acc0634b00fa4668ce6b4dee98621c0f70d803226c323753698c5ce Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.591294 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9r2kr"] Nov 24 06:57:30 crc kubenswrapper[4809]: W1124 06:57:30.603521 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c707d1d_d2c2_42db_82b9_afe8fde54914.slice/crio-8f7718e2a085edd75d56c3555f57acab01e0fb29524f3c74b10aad7f06cdde4b WatchSource:0}: Error finding container 8f7718e2a085edd75d56c3555f57acab01e0fb29524f3c74b10aad7f06cdde4b: Status 404 returned error can't find the container with id 8f7718e2a085edd75d56c3555f57acab01e0fb29524f3c74b10aad7f06cdde4b Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.686202 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7rrb4"] Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.688908 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.689023 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 06:57:30 crc kubenswrapper[4809]: W1124 06:57:30.697425 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc385307d_febf_4fe4_bdf6_51224328ea10.slice/crio-0858b5e27e106a15a0a62bbb489b55fc0200aecbc20568a1f6461bbec7852d1c WatchSource:0}: Error finding container 0858b5e27e106a15a0a62bbb489b55fc0200aecbc20568a1f6461bbec7852d1c: Status 404 returned error can't find the container with id 0858b5e27e106a15a0a62bbb489b55fc0200aecbc20568a1f6461bbec7852d1c Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.792436 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.792509 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.792860 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.814346 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.876046 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 06:57:30 crc kubenswrapper[4809]: I1124 06:57:30.907642 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.033809 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wsc5l"] Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.049803 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wsc5l"] Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.049898 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wsc5l" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.053409 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.065294 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" event={"ID":"4af1418f-a56e-4f26-a800-3b402b7da053","Type":"ContainerStarted","Data":"956426f3fe1032f711950ed48c6b6cff0626cabfb6e09ab8d7118943ffdfcd9e"} Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.065329 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" event={"ID":"4af1418f-a56e-4f26-a800-3b402b7da053","Type":"ContainerStarted","Data":"5921c15b8acc0634b00fa4668ce6b4dee98621c0f70d803226c323753698c5ce"} Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.065451 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.070390 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-bkt2p" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.070386 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-bkt2p" event={"ID":"476eceb4-f8fa-49c1-9015-5e067998af80","Type":"ContainerDied","Data":"c6c584b740382798cf36175787ddaa6614d5e0fb7719085e11c7f57f51e77a99"} Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.070497 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6c584b740382798cf36175787ddaa6614d5e0fb7719085e11c7f57f51e77a99" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.079162 4809 generic.go:334] "Generic (PLEG): container finished" podID="9c707d1d-d2c2-42db-82b9-afe8fde54914" containerID="56f4369c77a19d497136aba158c1b1e368e3fc74937c1cbbc1f4a5e5ea52a22d" exitCode=0 Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.079336 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9r2kr" event={"ID":"9c707d1d-d2c2-42db-82b9-afe8fde54914","Type":"ContainerDied","Data":"56f4369c77a19d497136aba158c1b1e368e3fc74937c1cbbc1f4a5e5ea52a22d"} Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.079387 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9r2kr" event={"ID":"9c707d1d-d2c2-42db-82b9-afe8fde54914","Type":"ContainerStarted","Data":"8f7718e2a085edd75d56c3555f57acab01e0fb29524f3c74b10aad7f06cdde4b"} Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.095047 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7rrb4" event={"ID":"c385307d-febf-4fe4-bdf6-51224328ea10","Type":"ContainerDied","Data":"19e5a66331aea317b795426c2fd47980fea970ffa24b10413d5d74c4ae8acf9b"} Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.095510 4809 generic.go:334] "Generic (PLEG): container finished" podID="c385307d-febf-4fe4-bdf6-51224328ea10" containerID="19e5a66331aea317b795426c2fd47980fea970ffa24b10413d5d74c4ae8acf9b" exitCode=0 Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.095587 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7rrb4" event={"ID":"c385307d-febf-4fe4-bdf6-51224328ea10","Type":"ContainerStarted","Data":"0858b5e27e106a15a0a62bbb489b55fc0200aecbc20568a1f6461bbec7852d1c"} Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.101614 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" podStartSLOduration=131.101590517 podStartE2EDuration="2m11.101590517s" podCreationTimestamp="2025-11-24 06:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:31.090328427 +0000 UTC m=+150.990920142" watchObservedRunningTime="2025-11-24 06:57:31.101590517 +0000 UTC m=+151.002182222" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.114297 4809 patch_prober.go:28] interesting pod/router-default-5444994796-r7k56 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:57:31 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Nov 24 06:57:31 crc kubenswrapper[4809]: [+]process-running ok Nov 24 06:57:31 crc kubenswrapper[4809]: healthz check failed Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.114373 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r7k56" podUID="9eab3672-3dd0-4af1-b96d-e8c5e99bb840" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.186334 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zhgn2" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.201387 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae7aaa0b-3da0-4764-808a-619dae2fe808-utilities\") pod \"redhat-operators-wsc5l\" (UID: \"ae7aaa0b-3da0-4764-808a-619dae2fe808\") " pod="openshift-marketplace/redhat-operators-wsc5l" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.201496 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6kpl\" (UniqueName: \"kubernetes.io/projected/ae7aaa0b-3da0-4764-808a-619dae2fe808-kube-api-access-j6kpl\") pod \"redhat-operators-wsc5l\" (UID: \"ae7aaa0b-3da0-4764-808a-619dae2fe808\") " pod="openshift-marketplace/redhat-operators-wsc5l" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.201555 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae7aaa0b-3da0-4764-808a-619dae2fe808-catalog-content\") pod \"redhat-operators-wsc5l\" (UID: \"ae7aaa0b-3da0-4764-808a-619dae2fe808\") " pod="openshift-marketplace/redhat-operators-wsc5l" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.302293 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae7aaa0b-3da0-4764-808a-619dae2fe808-utilities\") pod \"redhat-operators-wsc5l\" (UID: \"ae7aaa0b-3da0-4764-808a-619dae2fe808\") " pod="openshift-marketplace/redhat-operators-wsc5l" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.302404 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6kpl\" (UniqueName: \"kubernetes.io/projected/ae7aaa0b-3da0-4764-808a-619dae2fe808-kube-api-access-j6kpl\") pod \"redhat-operators-wsc5l\" (UID: \"ae7aaa0b-3da0-4764-808a-619dae2fe808\") " pod="openshift-marketplace/redhat-operators-wsc5l" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.302445 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae7aaa0b-3da0-4764-808a-619dae2fe808-catalog-content\") pod \"redhat-operators-wsc5l\" (UID: \"ae7aaa0b-3da0-4764-808a-619dae2fe808\") " pod="openshift-marketplace/redhat-operators-wsc5l" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.302853 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae7aaa0b-3da0-4764-808a-619dae2fe808-utilities\") pod \"redhat-operators-wsc5l\" (UID: \"ae7aaa0b-3da0-4764-808a-619dae2fe808\") " pod="openshift-marketplace/redhat-operators-wsc5l" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.302985 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae7aaa0b-3da0-4764-808a-619dae2fe808-catalog-content\") pod \"redhat-operators-wsc5l\" (UID: \"ae7aaa0b-3da0-4764-808a-619dae2fe808\") " pod="openshift-marketplace/redhat-operators-wsc5l" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.327444 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6kpl\" (UniqueName: \"kubernetes.io/projected/ae7aaa0b-3da0-4764-808a-619dae2fe808-kube-api-access-j6kpl\") pod \"redhat-operators-wsc5l\" (UID: \"ae7aaa0b-3da0-4764-808a-619dae2fe808\") " pod="openshift-marketplace/redhat-operators-wsc5l" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.371170 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.380905 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wsc5l" Nov 24 06:57:31 crc kubenswrapper[4809]: W1124 06:57:31.382184 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod2f0b7e2d_ada0_4191_b78b_fb0dcdd5c7aa.slice/crio-df60c0d3f2290a5287e04a11e7e2501548bab72ee20beb7cb21612fcd3fbc726 WatchSource:0}: Error finding container df60c0d3f2290a5287e04a11e7e2501548bab72ee20beb7cb21612fcd3fbc726: Status 404 returned error can't find the container with id df60c0d3f2290a5287e04a11e7e2501548bab72ee20beb7cb21612fcd3fbc726 Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.440002 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kppbx"] Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.441738 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kppbx" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.447234 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kppbx"] Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.612765 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jv2vk\" (UniqueName: \"kubernetes.io/projected/752560ef-f411-46b8-b3e7-d061cdc1e68b-kube-api-access-jv2vk\") pod \"redhat-operators-kppbx\" (UID: \"752560ef-f411-46b8-b3e7-d061cdc1e68b\") " pod="openshift-marketplace/redhat-operators-kppbx" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.613183 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/752560ef-f411-46b8-b3e7-d061cdc1e68b-catalog-content\") pod \"redhat-operators-kppbx\" (UID: \"752560ef-f411-46b8-b3e7-d061cdc1e68b\") " pod="openshift-marketplace/redhat-operators-kppbx" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.613273 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/752560ef-f411-46b8-b3e7-d061cdc1e68b-utilities\") pod \"redhat-operators-kppbx\" (UID: \"752560ef-f411-46b8-b3e7-d061cdc1e68b\") " pod="openshift-marketplace/redhat-operators-kppbx" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.643180 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.643257 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.648129 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.648186 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.652552 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.654370 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.665823 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wsc5l"] Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.678633 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-lkdj2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.678681 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lkdj2" podUID="b6237cc7-ba93-45a6-b560-d7916abf71f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.679301 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-lkdj2 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.679349 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-lkdj2" podUID="b6237cc7-ba93-45a6-b560-d7916abf71f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Nov 24 06:57:31 crc kubenswrapper[4809]: W1124 06:57:31.690421 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae7aaa0b_3da0_4764_808a_619dae2fe808.slice/crio-b4355799703c087442c717fe6dec81edcb8f451852047a00f6d985e696796533 WatchSource:0}: Error finding container b4355799703c087442c717fe6dec81edcb8f451852047a00f6d985e696796533: Status 404 returned error can't find the container with id b4355799703c087442c717fe6dec81edcb8f451852047a00f6d985e696796533 Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.714539 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/752560ef-f411-46b8-b3e7-d061cdc1e68b-utilities\") pod \"redhat-operators-kppbx\" (UID: \"752560ef-f411-46b8-b3e7-d061cdc1e68b\") " pod="openshift-marketplace/redhat-operators-kppbx" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.714622 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jv2vk\" (UniqueName: \"kubernetes.io/projected/752560ef-f411-46b8-b3e7-d061cdc1e68b-kube-api-access-jv2vk\") pod \"redhat-operators-kppbx\" (UID: \"752560ef-f411-46b8-b3e7-d061cdc1e68b\") " pod="openshift-marketplace/redhat-operators-kppbx" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.714653 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/752560ef-f411-46b8-b3e7-d061cdc1e68b-catalog-content\") pod \"redhat-operators-kppbx\" (UID: \"752560ef-f411-46b8-b3e7-d061cdc1e68b\") " pod="openshift-marketplace/redhat-operators-kppbx" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.716683 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/752560ef-f411-46b8-b3e7-d061cdc1e68b-utilities\") pod \"redhat-operators-kppbx\" (UID: \"752560ef-f411-46b8-b3e7-d061cdc1e68b\") " pod="openshift-marketplace/redhat-operators-kppbx" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.716703 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/752560ef-f411-46b8-b3e7-d061cdc1e68b-catalog-content\") pod \"redhat-operators-kppbx\" (UID: \"752560ef-f411-46b8-b3e7-d061cdc1e68b\") " pod="openshift-marketplace/redhat-operators-kppbx" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.735296 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jv2vk\" (UniqueName: \"kubernetes.io/projected/752560ef-f411-46b8-b3e7-d061cdc1e68b-kube-api-access-jv2vk\") pod \"redhat-operators-kppbx\" (UID: \"752560ef-f411-46b8-b3e7-d061cdc1e68b\") " pod="openshift-marketplace/redhat-operators-kppbx" Nov 24 06:57:31 crc kubenswrapper[4809]: I1124 06:57:31.795202 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kppbx" Nov 24 06:57:32 crc kubenswrapper[4809]: I1124 06:57:32.026113 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kppbx"] Nov 24 06:57:32 crc kubenswrapper[4809]: W1124 06:57:32.091404 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod752560ef_f411_46b8_b3e7_d061cdc1e68b.slice/crio-fedb1888f481d996552a6c9ff34fddd0eab8dbaeb5aaa1b825fb391604876884 WatchSource:0}: Error finding container fedb1888f481d996552a6c9ff34fddd0eab8dbaeb5aaa1b825fb391604876884: Status 404 returned error can't find the container with id fedb1888f481d996552a6c9ff34fddd0eab8dbaeb5aaa1b825fb391604876884 Nov 24 06:57:32 crc kubenswrapper[4809]: I1124 06:57:32.111149 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-r7k56" Nov 24 06:57:32 crc kubenswrapper[4809]: I1124 06:57:32.114297 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:32 crc kubenswrapper[4809]: I1124 06:57:32.115570 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:32 crc kubenswrapper[4809]: I1124 06:57:32.116689 4809 patch_prober.go:28] interesting pod/router-default-5444994796-r7k56 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:57:32 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Nov 24 06:57:32 crc kubenswrapper[4809]: [+]process-running ok Nov 24 06:57:32 crc kubenswrapper[4809]: healthz check failed Nov 24 06:57:32 crc kubenswrapper[4809]: I1124 06:57:32.116743 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r7k56" podUID="9eab3672-3dd0-4af1-b96d-e8c5e99bb840" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:57:32 crc kubenswrapper[4809]: I1124 06:57:32.117770 4809 patch_prober.go:28] interesting pod/console-f9d7485db-8r5vw container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Nov 24 06:57:32 crc kubenswrapper[4809]: I1124 06:57:32.117826 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-8r5vw" podUID="2fe229a6-e9e2-40b3-a4d7-0780c4f5c884" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Nov 24 06:57:32 crc kubenswrapper[4809]: I1124 06:57:32.118600 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa","Type":"ContainerStarted","Data":"3cf029776055a4f4a82cc9d56df4fd9a1ac3ae5988589e4ac6d718f898ceb3bd"} Nov 24 06:57:32 crc kubenswrapper[4809]: I1124 06:57:32.118673 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa","Type":"ContainerStarted","Data":"df60c0d3f2290a5287e04a11e7e2501548bab72ee20beb7cb21612fcd3fbc726"} Nov 24 06:57:32 crc kubenswrapper[4809]: I1124 06:57:32.128897 4809 generic.go:334] "Generic (PLEG): container finished" podID="ae7aaa0b-3da0-4764-808a-619dae2fe808" containerID="9275f80fce7ee8d0fcacf680bf7f71ee2ae6f6be736e2da04beaeb19bc7d332b" exitCode=0 Nov 24 06:57:32 crc kubenswrapper[4809]: I1124 06:57:32.129049 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsc5l" event={"ID":"ae7aaa0b-3da0-4764-808a-619dae2fe808","Type":"ContainerDied","Data":"9275f80fce7ee8d0fcacf680bf7f71ee2ae6f6be736e2da04beaeb19bc7d332b"} Nov 24 06:57:32 crc kubenswrapper[4809]: I1124 06:57:32.129435 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsc5l" event={"ID":"ae7aaa0b-3da0-4764-808a-619dae2fe808","Type":"ContainerStarted","Data":"b4355799703c087442c717fe6dec81edcb8f451852047a00f6d985e696796533"} Nov 24 06:57:32 crc kubenswrapper[4809]: I1124 06:57:32.136443 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.136425765 podStartE2EDuration="2.136425765s" podCreationTimestamp="2025-11-24 06:57:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:32.133828619 +0000 UTC m=+152.034420324" watchObservedRunningTime="2025-11-24 06:57:32.136425765 +0000 UTC m=+152.037017470" Nov 24 06:57:32 crc kubenswrapper[4809]: I1124 06:57:32.137390 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kppbx" event={"ID":"752560ef-f411-46b8-b3e7-d061cdc1e68b","Type":"ContainerStarted","Data":"fedb1888f481d996552a6c9ff34fddd0eab8dbaeb5aaa1b825fb391604876884"} Nov 24 06:57:32 crc kubenswrapper[4809]: I1124 06:57:32.141578 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-88wkc" Nov 24 06:57:32 crc kubenswrapper[4809]: I1124 06:57:32.151099 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-88j9l" Nov 24 06:57:32 crc kubenswrapper[4809]: I1124 06:57:32.620094 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" Nov 24 06:57:33 crc kubenswrapper[4809]: I1124 06:57:33.110517 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-r7k56" Nov 24 06:57:33 crc kubenswrapper[4809]: I1124 06:57:33.153070 4809 generic.go:334] "Generic (PLEG): container finished" podID="752560ef-f411-46b8-b3e7-d061cdc1e68b" containerID="a44d1727650fb8334c7dc28b2da024870ca16c10ea19c90c4c131989a11386f7" exitCode=0 Nov 24 06:57:33 crc kubenswrapper[4809]: I1124 06:57:33.153111 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kppbx" event={"ID":"752560ef-f411-46b8-b3e7-d061cdc1e68b","Type":"ContainerDied","Data":"a44d1727650fb8334c7dc28b2da024870ca16c10ea19c90c4c131989a11386f7"} Nov 24 06:57:33 crc kubenswrapper[4809]: I1124 06:57:33.156559 4809 generic.go:334] "Generic (PLEG): container finished" podID="2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa" containerID="3cf029776055a4f4a82cc9d56df4fd9a1ac3ae5988589e4ac6d718f898ceb3bd" exitCode=0 Nov 24 06:57:33 crc kubenswrapper[4809]: I1124 06:57:33.156642 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa","Type":"ContainerDied","Data":"3cf029776055a4f4a82cc9d56df4fd9a1ac3ae5988589e4ac6d718f898ceb3bd"} Nov 24 06:57:33 crc kubenswrapper[4809]: I1124 06:57:33.160719 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-r7k56" Nov 24 06:57:33 crc kubenswrapper[4809]: I1124 06:57:33.290861 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 06:57:33 crc kubenswrapper[4809]: I1124 06:57:33.292149 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 06:57:33 crc kubenswrapper[4809]: I1124 06:57:33.303083 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 24 06:57:33 crc kubenswrapper[4809]: I1124 06:57:33.307252 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 24 06:57:33 crc kubenswrapper[4809]: I1124 06:57:33.307452 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 06:57:33 crc kubenswrapper[4809]: I1124 06:57:33.366607 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/442d2bfc-3172-4bd8-bfb6-f14668b8c971-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"442d2bfc-3172-4bd8-bfb6-f14668b8c971\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 06:57:33 crc kubenswrapper[4809]: I1124 06:57:33.366739 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/442d2bfc-3172-4bd8-bfb6-f14668b8c971-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"442d2bfc-3172-4bd8-bfb6-f14668b8c971\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 06:57:33 crc kubenswrapper[4809]: I1124 06:57:33.467892 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/442d2bfc-3172-4bd8-bfb6-f14668b8c971-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"442d2bfc-3172-4bd8-bfb6-f14668b8c971\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 06:57:33 crc kubenswrapper[4809]: I1124 06:57:33.468229 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/442d2bfc-3172-4bd8-bfb6-f14668b8c971-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"442d2bfc-3172-4bd8-bfb6-f14668b8c971\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 06:57:33 crc kubenswrapper[4809]: I1124 06:57:33.468345 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/442d2bfc-3172-4bd8-bfb6-f14668b8c971-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"442d2bfc-3172-4bd8-bfb6-f14668b8c971\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 06:57:33 crc kubenswrapper[4809]: I1124 06:57:33.501266 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/442d2bfc-3172-4bd8-bfb6-f14668b8c971-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"442d2bfc-3172-4bd8-bfb6-f14668b8c971\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 06:57:33 crc kubenswrapper[4809]: I1124 06:57:33.625559 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 06:57:33 crc kubenswrapper[4809]: I1124 06:57:33.975874 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 06:57:34 crc kubenswrapper[4809]: W1124 06:57:34.049107 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod442d2bfc_3172_4bd8_bfb6_f14668b8c971.slice/crio-f07f863e741d2897e0eb308cb6486d1adbf5439a9a127b2c61a1805990f8d402 WatchSource:0}: Error finding container f07f863e741d2897e0eb308cb6486d1adbf5439a9a127b2c61a1805990f8d402: Status 404 returned error can't find the container with id f07f863e741d2897e0eb308cb6486d1adbf5439a9a127b2c61a1805990f8d402 Nov 24 06:57:34 crc kubenswrapper[4809]: I1124 06:57:34.166568 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"442d2bfc-3172-4bd8-bfb6-f14668b8c971","Type":"ContainerStarted","Data":"f07f863e741d2897e0eb308cb6486d1adbf5439a9a127b2c61a1805990f8d402"} Nov 24 06:57:34 crc kubenswrapper[4809]: I1124 06:57:34.400644 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 06:57:34 crc kubenswrapper[4809]: I1124 06:57:34.482690 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa-kube-api-access\") pod \"2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa\" (UID: \"2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa\") " Nov 24 06:57:34 crc kubenswrapper[4809]: I1124 06:57:34.483144 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa-kubelet-dir\") pod \"2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa\" (UID: \"2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa\") " Nov 24 06:57:34 crc kubenswrapper[4809]: I1124 06:57:34.483224 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa" (UID: "2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 06:57:34 crc kubenswrapper[4809]: I1124 06:57:34.483574 4809 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 06:57:34 crc kubenswrapper[4809]: I1124 06:57:34.488329 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa" (UID: "2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:57:34 crc kubenswrapper[4809]: I1124 06:57:34.585169 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 06:57:35 crc kubenswrapper[4809]: I1124 06:57:35.177729 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa","Type":"ContainerDied","Data":"df60c0d3f2290a5287e04a11e7e2501548bab72ee20beb7cb21612fcd3fbc726"} Nov 24 06:57:35 crc kubenswrapper[4809]: I1124 06:57:35.177766 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df60c0d3f2290a5287e04a11e7e2501548bab72ee20beb7cb21612fcd3fbc726" Nov 24 06:57:35 crc kubenswrapper[4809]: I1124 06:57:35.177818 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 06:57:35 crc kubenswrapper[4809]: I1124 06:57:35.183945 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"442d2bfc-3172-4bd8-bfb6-f14668b8c971","Type":"ContainerStarted","Data":"b4148f95f9f1687e50172153343c555b49a85752d3d42ad36f13dca03f725787"} Nov 24 06:57:35 crc kubenswrapper[4809]: I1124 06:57:35.203780 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.203758355 podStartE2EDuration="2.203758355s" podCreationTimestamp="2025-11-24 06:57:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:57:35.195177303 +0000 UTC m=+155.095769008" watchObservedRunningTime="2025-11-24 06:57:35.203758355 +0000 UTC m=+155.104350060" Nov 24 06:57:36 crc kubenswrapper[4809]: I1124 06:57:36.219402 4809 generic.go:334] "Generic (PLEG): container finished" podID="442d2bfc-3172-4bd8-bfb6-f14668b8c971" containerID="b4148f95f9f1687e50172153343c555b49a85752d3d42ad36f13dca03f725787" exitCode=0 Nov 24 06:57:36 crc kubenswrapper[4809]: I1124 06:57:36.219463 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"442d2bfc-3172-4bd8-bfb6-f14668b8c971","Type":"ContainerDied","Data":"b4148f95f9f1687e50172153343c555b49a85752d3d42ad36f13dca03f725787"} Nov 24 06:57:37 crc kubenswrapper[4809]: I1124 06:57:37.706145 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-wlshf" Nov 24 06:57:41 crc kubenswrapper[4809]: I1124 06:57:41.688449 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-lkdj2" Nov 24 06:57:42 crc kubenswrapper[4809]: I1124 06:57:42.115072 4809 patch_prober.go:28] interesting pod/console-f9d7485db-8r5vw container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Nov 24 06:57:42 crc kubenswrapper[4809]: I1124 06:57:42.115346 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-8r5vw" podUID="2fe229a6-e9e2-40b3-a4d7-0780c4f5c884" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Nov 24 06:57:42 crc kubenswrapper[4809]: I1124 06:57:42.505990 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs\") pod \"network-metrics-daemon-krmz8\" (UID: \"35b8615e-faa5-47b4-8c22-609d5fda9590\") " pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:57:42 crc kubenswrapper[4809]: I1124 06:57:42.512448 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/35b8615e-faa5-47b4-8c22-609d5fda9590-metrics-certs\") pod \"network-metrics-daemon-krmz8\" (UID: \"35b8615e-faa5-47b4-8c22-609d5fda9590\") " pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:57:42 crc kubenswrapper[4809]: I1124 06:57:42.623887 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krmz8" Nov 24 06:57:45 crc kubenswrapper[4809]: I1124 06:57:45.006372 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 06:57:45 crc kubenswrapper[4809]: I1124 06:57:45.141830 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/442d2bfc-3172-4bd8-bfb6-f14668b8c971-kubelet-dir\") pod \"442d2bfc-3172-4bd8-bfb6-f14668b8c971\" (UID: \"442d2bfc-3172-4bd8-bfb6-f14668b8c971\") " Nov 24 06:57:45 crc kubenswrapper[4809]: I1124 06:57:45.141895 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/442d2bfc-3172-4bd8-bfb6-f14668b8c971-kube-api-access\") pod \"442d2bfc-3172-4bd8-bfb6-f14668b8c971\" (UID: \"442d2bfc-3172-4bd8-bfb6-f14668b8c971\") " Nov 24 06:57:45 crc kubenswrapper[4809]: I1124 06:57:45.141995 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/442d2bfc-3172-4bd8-bfb6-f14668b8c971-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "442d2bfc-3172-4bd8-bfb6-f14668b8c971" (UID: "442d2bfc-3172-4bd8-bfb6-f14668b8c971"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 06:57:45 crc kubenswrapper[4809]: I1124 06:57:45.142244 4809 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/442d2bfc-3172-4bd8-bfb6-f14668b8c971-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 06:57:45 crc kubenswrapper[4809]: I1124 06:57:45.160776 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/442d2bfc-3172-4bd8-bfb6-f14668b8c971-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "442d2bfc-3172-4bd8-bfb6-f14668b8c971" (UID: "442d2bfc-3172-4bd8-bfb6-f14668b8c971"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:57:45 crc kubenswrapper[4809]: I1124 06:57:45.243780 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/442d2bfc-3172-4bd8-bfb6-f14668b8c971-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 06:57:45 crc kubenswrapper[4809]: I1124 06:57:45.287717 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"442d2bfc-3172-4bd8-bfb6-f14668b8c971","Type":"ContainerDied","Data":"f07f863e741d2897e0eb308cb6486d1adbf5439a9a127b2c61a1805990f8d402"} Nov 24 06:57:45 crc kubenswrapper[4809]: I1124 06:57:45.287755 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f07f863e741d2897e0eb308cb6486d1adbf5439a9a127b2c61a1805990f8d402" Nov 24 06:57:45 crc kubenswrapper[4809]: I1124 06:57:45.287748 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 06:57:48 crc kubenswrapper[4809]: I1124 06:57:48.043619 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 06:57:48 crc kubenswrapper[4809]: I1124 06:57:48.043762 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 06:57:50 crc kubenswrapper[4809]: I1124 06:57:50.343436 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 06:57:52 crc kubenswrapper[4809]: I1124 06:57:52.118679 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:57:52 crc kubenswrapper[4809]: I1124 06:57:52.122414 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 06:58:01 crc kubenswrapper[4809]: E1124 06:58:01.781378 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 24 06:58:01 crc kubenswrapper[4809]: E1124 06:58:01.781658 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-58z8t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-llxlp_openshift-marketplace(f8951296-b323-4ac3-9e01-963b60ff998c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 06:58:01 crc kubenswrapper[4809]: E1124 06:58:01.783056 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-llxlp" podUID="f8951296-b323-4ac3-9e01-963b60ff998c" Nov 24 06:58:01 crc kubenswrapper[4809]: E1124 06:58:01.811600 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 24 06:58:01 crc kubenswrapper[4809]: E1124 06:58:01.811983 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kcbrr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-7rrb4_openshift-marketplace(c385307d-febf-4fe4-bdf6-51224328ea10): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 06:58:01 crc kubenswrapper[4809]: E1124 06:58:01.814674 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-7rrb4" podUID="c385307d-febf-4fe4-bdf6-51224328ea10" Nov 24 06:58:02 crc kubenswrapper[4809]: I1124 06:58:02.573471 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5w6mp" Nov 24 06:58:04 crc kubenswrapper[4809]: E1124 06:58:04.686167 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-llxlp" podUID="f8951296-b323-4ac3-9e01-963b60ff998c" Nov 24 06:58:04 crc kubenswrapper[4809]: E1124 06:58:04.686237 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7rrb4" podUID="c385307d-febf-4fe4-bdf6-51224328ea10" Nov 24 06:58:04 crc kubenswrapper[4809]: E1124 06:58:04.769341 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 24 06:58:04 crc kubenswrapper[4809]: E1124 06:58:04.769891 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j6kpl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-wsc5l_openshift-marketplace(ae7aaa0b-3da0-4764-808a-619dae2fe808): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 06:58:04 crc kubenswrapper[4809]: E1124 06:58:04.771590 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-wsc5l" podUID="ae7aaa0b-3da0-4764-808a-619dae2fe808" Nov 24 06:58:04 crc kubenswrapper[4809]: E1124 06:58:04.835050 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 24 06:58:04 crc kubenswrapper[4809]: E1124 06:58:04.835192 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-drtpr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-9r2kr_openshift-marketplace(9c707d1d-d2c2-42db-82b9-afe8fde54914): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 06:58:04 crc kubenswrapper[4809]: E1124 06:58:04.836344 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-9r2kr" podUID="9c707d1d-d2c2-42db-82b9-afe8fde54914" Nov 24 06:58:04 crc kubenswrapper[4809]: E1124 06:58:04.836717 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 24 06:58:04 crc kubenswrapper[4809]: E1124 06:58:04.836859 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hk262,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-pj46g_openshift-marketplace(1383ed81-5c2a-4416-8c67-05e4881db7be): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 06:58:04 crc kubenswrapper[4809]: E1124 06:58:04.838660 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-pj46g" podUID="1383ed81-5c2a-4416-8c67-05e4881db7be" Nov 24 06:58:04 crc kubenswrapper[4809]: E1124 06:58:04.843609 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 24 06:58:04 crc kubenswrapper[4809]: E1124 06:58:04.843763 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vm75s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-9ttsc_openshift-marketplace(8c69d61e-4b0d-4388-aff9-9571d2a9c16c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 06:58:04 crc kubenswrapper[4809]: E1124 06:58:04.844911 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-9ttsc" podUID="8c69d61e-4b0d-4388-aff9-9571d2a9c16c" Nov 24 06:58:05 crc kubenswrapper[4809]: I1124 06:58:05.150107 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-krmz8"] Nov 24 06:58:05 crc kubenswrapper[4809]: W1124 06:58:05.159379 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod35b8615e_faa5_47b4_8c22_609d5fda9590.slice/crio-0ebced32d6c65f8c75322077627f1cf9bbbb6b07d374864f9592745e16bd6b12 WatchSource:0}: Error finding container 0ebced32d6c65f8c75322077627f1cf9bbbb6b07d374864f9592745e16bd6b12: Status 404 returned error can't find the container with id 0ebced32d6c65f8c75322077627f1cf9bbbb6b07d374864f9592745e16bd6b12 Nov 24 06:58:05 crc kubenswrapper[4809]: I1124 06:58:05.386811 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-krmz8" event={"ID":"35b8615e-faa5-47b4-8c22-609d5fda9590","Type":"ContainerStarted","Data":"a885da6266f84bd603d59cfb837a532b9c030b6d60533b9cc86684223cc1750c"} Nov 24 06:58:05 crc kubenswrapper[4809]: I1124 06:58:05.387160 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-krmz8" event={"ID":"35b8615e-faa5-47b4-8c22-609d5fda9590","Type":"ContainerStarted","Data":"0ebced32d6c65f8c75322077627f1cf9bbbb6b07d374864f9592745e16bd6b12"} Nov 24 06:58:05 crc kubenswrapper[4809]: I1124 06:58:05.389665 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kppbx" event={"ID":"752560ef-f411-46b8-b3e7-d061cdc1e68b","Type":"ContainerStarted","Data":"ec44bd017adf72220beb0ffb97c95012466c69749ff09d0b8b1d316bf1a554a3"} Nov 24 06:58:05 crc kubenswrapper[4809]: I1124 06:58:05.392714 4809 generic.go:334] "Generic (PLEG): container finished" podID="fb64a807-42bd-4297-aa31-52a142255d89" containerID="d7b8c69c00a14ca92bf215ec395a6cb2d9f86b8ed9407f4d5b2b8073ca2f1b43" exitCode=0 Nov 24 06:58:05 crc kubenswrapper[4809]: I1124 06:58:05.392843 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bnt6m" event={"ID":"fb64a807-42bd-4297-aa31-52a142255d89","Type":"ContainerDied","Data":"d7b8c69c00a14ca92bf215ec395a6cb2d9f86b8ed9407f4d5b2b8073ca2f1b43"} Nov 24 06:58:05 crc kubenswrapper[4809]: E1124 06:58:05.396241 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-9r2kr" podUID="9c707d1d-d2c2-42db-82b9-afe8fde54914" Nov 24 06:58:05 crc kubenswrapper[4809]: E1124 06:58:05.396405 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-9ttsc" podUID="8c69d61e-4b0d-4388-aff9-9571d2a9c16c" Nov 24 06:58:05 crc kubenswrapper[4809]: E1124 06:58:05.397466 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-pj46g" podUID="1383ed81-5c2a-4416-8c67-05e4881db7be" Nov 24 06:58:05 crc kubenswrapper[4809]: E1124 06:58:05.398010 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-wsc5l" podUID="ae7aaa0b-3da0-4764-808a-619dae2fe808" Nov 24 06:58:06 crc kubenswrapper[4809]: I1124 06:58:06.410434 4809 generic.go:334] "Generic (PLEG): container finished" podID="752560ef-f411-46b8-b3e7-d061cdc1e68b" containerID="ec44bd017adf72220beb0ffb97c95012466c69749ff09d0b8b1d316bf1a554a3" exitCode=0 Nov 24 06:58:06 crc kubenswrapper[4809]: I1124 06:58:06.410501 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kppbx" event={"ID":"752560ef-f411-46b8-b3e7-d061cdc1e68b","Type":"ContainerDied","Data":"ec44bd017adf72220beb0ffb97c95012466c69749ff09d0b8b1d316bf1a554a3"} Nov 24 06:58:06 crc kubenswrapper[4809]: I1124 06:58:06.419353 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bnt6m" event={"ID":"fb64a807-42bd-4297-aa31-52a142255d89","Type":"ContainerStarted","Data":"126b6641e621fde2b5a56b5ad730c6e7322c5ddefb916a17550cb1ecc1871684"} Nov 24 06:58:06 crc kubenswrapper[4809]: I1124 06:58:06.421391 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-krmz8" event={"ID":"35b8615e-faa5-47b4-8c22-609d5fda9590","Type":"ContainerStarted","Data":"9c3ba52b77655cce238da3ac34892552be33934ee917458c873fd2aaca1e014b"} Nov 24 06:58:06 crc kubenswrapper[4809]: I1124 06:58:06.459069 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bnt6m" podStartSLOduration=2.695162447 podStartE2EDuration="38.459047189s" podCreationTimestamp="2025-11-24 06:57:28 +0000 UTC" firstStartedPulling="2025-11-24 06:57:30.040595612 +0000 UTC m=+149.941187317" lastFinishedPulling="2025-11-24 06:58:05.804480354 +0000 UTC m=+185.705072059" observedRunningTime="2025-11-24 06:58:06.457862783 +0000 UTC m=+186.358454488" watchObservedRunningTime="2025-11-24 06:58:06.459047189 +0000 UTC m=+186.359638924" Nov 24 06:58:07 crc kubenswrapper[4809]: I1124 06:58:07.428179 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kppbx" event={"ID":"752560ef-f411-46b8-b3e7-d061cdc1e68b","Type":"ContainerStarted","Data":"3c63c013b1d02bca0030cc98821d18401b6f6106e814927ac3adfe3c7a0e09c3"} Nov 24 06:58:07 crc kubenswrapper[4809]: I1124 06:58:07.449399 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-krmz8" podStartSLOduration=166.449381909 podStartE2EDuration="2m46.449381909s" podCreationTimestamp="2025-11-24 06:55:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:58:06.476545002 +0000 UTC m=+186.377136727" watchObservedRunningTime="2025-11-24 06:58:07.449381909 +0000 UTC m=+187.349973614" Nov 24 06:58:07 crc kubenswrapper[4809]: I1124 06:58:07.451229 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kppbx" podStartSLOduration=2.821380087 podStartE2EDuration="36.451222124s" podCreationTimestamp="2025-11-24 06:57:31 +0000 UTC" firstStartedPulling="2025-11-24 06:57:33.164851274 +0000 UTC m=+153.065442979" lastFinishedPulling="2025-11-24 06:58:06.794693311 +0000 UTC m=+186.695285016" observedRunningTime="2025-11-24 06:58:07.447783092 +0000 UTC m=+187.348374797" watchObservedRunningTime="2025-11-24 06:58:07.451222124 +0000 UTC m=+187.351813829" Nov 24 06:58:08 crc kubenswrapper[4809]: I1124 06:58:08.321594 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:58:08 crc kubenswrapper[4809]: I1124 06:58:08.365311 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bnt6m" Nov 24 06:58:08 crc kubenswrapper[4809]: I1124 06:58:08.365361 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bnt6m" Nov 24 06:58:08 crc kubenswrapper[4809]: I1124 06:58:08.517249 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bnt6m" Nov 24 06:58:11 crc kubenswrapper[4809]: I1124 06:58:11.796093 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kppbx" Nov 24 06:58:11 crc kubenswrapper[4809]: I1124 06:58:11.797056 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kppbx" Nov 24 06:58:12 crc kubenswrapper[4809]: I1124 06:58:12.834320 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kppbx" podUID="752560ef-f411-46b8-b3e7-d061cdc1e68b" containerName="registry-server" probeResult="failure" output=< Nov 24 06:58:12 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Nov 24 06:58:12 crc kubenswrapper[4809]: > Nov 24 06:58:18 crc kubenswrapper[4809]: I1124 06:58:18.043280 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 06:58:18 crc kubenswrapper[4809]: I1124 06:58:18.043867 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 06:58:18 crc kubenswrapper[4809]: I1124 06:58:18.403903 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bnt6m" Nov 24 06:58:19 crc kubenswrapper[4809]: I1124 06:58:19.319501 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bnt6m"] Nov 24 06:58:19 crc kubenswrapper[4809]: I1124 06:58:19.319761 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bnt6m" podUID="fb64a807-42bd-4297-aa31-52a142255d89" containerName="registry-server" containerID="cri-o://126b6641e621fde2b5a56b5ad730c6e7322c5ddefb916a17550cb1ecc1871684" gracePeriod=2 Nov 24 06:58:19 crc kubenswrapper[4809]: I1124 06:58:19.508315 4809 generic.go:334] "Generic (PLEG): container finished" podID="fb64a807-42bd-4297-aa31-52a142255d89" containerID="126b6641e621fde2b5a56b5ad730c6e7322c5ddefb916a17550cb1ecc1871684" exitCode=0 Nov 24 06:58:19 crc kubenswrapper[4809]: I1124 06:58:19.508467 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bnt6m" event={"ID":"fb64a807-42bd-4297-aa31-52a142255d89","Type":"ContainerDied","Data":"126b6641e621fde2b5a56b5ad730c6e7322c5ddefb916a17550cb1ecc1871684"} Nov 24 06:58:19 crc kubenswrapper[4809]: I1124 06:58:19.748371 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bnt6m" Nov 24 06:58:19 crc kubenswrapper[4809]: I1124 06:58:19.818292 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb64a807-42bd-4297-aa31-52a142255d89-utilities\") pod \"fb64a807-42bd-4297-aa31-52a142255d89\" (UID: \"fb64a807-42bd-4297-aa31-52a142255d89\") " Nov 24 06:58:19 crc kubenswrapper[4809]: I1124 06:58:19.818705 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb64a807-42bd-4297-aa31-52a142255d89-catalog-content\") pod \"fb64a807-42bd-4297-aa31-52a142255d89\" (UID: \"fb64a807-42bd-4297-aa31-52a142255d89\") " Nov 24 06:58:19 crc kubenswrapper[4809]: I1124 06:58:19.818737 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qglxv\" (UniqueName: \"kubernetes.io/projected/fb64a807-42bd-4297-aa31-52a142255d89-kube-api-access-qglxv\") pod \"fb64a807-42bd-4297-aa31-52a142255d89\" (UID: \"fb64a807-42bd-4297-aa31-52a142255d89\") " Nov 24 06:58:19 crc kubenswrapper[4809]: I1124 06:58:19.819098 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb64a807-42bd-4297-aa31-52a142255d89-utilities" (OuterVolumeSpecName: "utilities") pod "fb64a807-42bd-4297-aa31-52a142255d89" (UID: "fb64a807-42bd-4297-aa31-52a142255d89"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:58:19 crc kubenswrapper[4809]: I1124 06:58:19.824140 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb64a807-42bd-4297-aa31-52a142255d89-kube-api-access-qglxv" (OuterVolumeSpecName: "kube-api-access-qglxv") pod "fb64a807-42bd-4297-aa31-52a142255d89" (UID: "fb64a807-42bd-4297-aa31-52a142255d89"). InnerVolumeSpecName "kube-api-access-qglxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:58:19 crc kubenswrapper[4809]: I1124 06:58:19.827556 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb64a807-42bd-4297-aa31-52a142255d89-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:19 crc kubenswrapper[4809]: I1124 06:58:19.827585 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qglxv\" (UniqueName: \"kubernetes.io/projected/fb64a807-42bd-4297-aa31-52a142255d89-kube-api-access-qglxv\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:19 crc kubenswrapper[4809]: I1124 06:58:19.873437 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb64a807-42bd-4297-aa31-52a142255d89-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fb64a807-42bd-4297-aa31-52a142255d89" (UID: "fb64a807-42bd-4297-aa31-52a142255d89"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:58:19 crc kubenswrapper[4809]: I1124 06:58:19.929525 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb64a807-42bd-4297-aa31-52a142255d89-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:20 crc kubenswrapper[4809]: I1124 06:58:20.516958 4809 generic.go:334] "Generic (PLEG): container finished" podID="c385307d-febf-4fe4-bdf6-51224328ea10" containerID="5c288b1bcd9d858a0d768a8816cee9c7a4ee05bc910ed2b73214a7af339b4ba9" exitCode=0 Nov 24 06:58:20 crc kubenswrapper[4809]: I1124 06:58:20.516996 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7rrb4" event={"ID":"c385307d-febf-4fe4-bdf6-51224328ea10","Type":"ContainerDied","Data":"5c288b1bcd9d858a0d768a8816cee9c7a4ee05bc910ed2b73214a7af339b4ba9"} Nov 24 06:58:20 crc kubenswrapper[4809]: I1124 06:58:20.519384 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bnt6m" Nov 24 06:58:20 crc kubenswrapper[4809]: I1124 06:58:20.519596 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bnt6m" event={"ID":"fb64a807-42bd-4297-aa31-52a142255d89","Type":"ContainerDied","Data":"b6b42b0a5363a81f8bb60d7c0456d5df702d747cff9b7e0ae92c835cb677debd"} Nov 24 06:58:20 crc kubenswrapper[4809]: I1124 06:58:20.519704 4809 scope.go:117] "RemoveContainer" containerID="126b6641e621fde2b5a56b5ad730c6e7322c5ddefb916a17550cb1ecc1871684" Nov 24 06:58:20 crc kubenswrapper[4809]: I1124 06:58:20.521989 4809 generic.go:334] "Generic (PLEG): container finished" podID="8c69d61e-4b0d-4388-aff9-9571d2a9c16c" containerID="c3c577bbcad29f4843a716708cc19b98cba12bbbc5ae59a46de75fd96c30b961" exitCode=0 Nov 24 06:58:20 crc kubenswrapper[4809]: I1124 06:58:20.522002 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9ttsc" event={"ID":"8c69d61e-4b0d-4388-aff9-9571d2a9c16c","Type":"ContainerDied","Data":"c3c577bbcad29f4843a716708cc19b98cba12bbbc5ae59a46de75fd96c30b961"} Nov 24 06:58:20 crc kubenswrapper[4809]: I1124 06:58:20.524772 4809 generic.go:334] "Generic (PLEG): container finished" podID="ae7aaa0b-3da0-4764-808a-619dae2fe808" containerID="22bcfb0f275251f85af8aadda4c68d69015b9f71e19a4c45b74c420b8838d835" exitCode=0 Nov 24 06:58:20 crc kubenswrapper[4809]: I1124 06:58:20.524822 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsc5l" event={"ID":"ae7aaa0b-3da0-4764-808a-619dae2fe808","Type":"ContainerDied","Data":"22bcfb0f275251f85af8aadda4c68d69015b9f71e19a4c45b74c420b8838d835"} Nov 24 06:58:20 crc kubenswrapper[4809]: I1124 06:58:20.548322 4809 scope.go:117] "RemoveContainer" containerID="d7b8c69c00a14ca92bf215ec395a6cb2d9f86b8ed9407f4d5b2b8073ca2f1b43" Nov 24 06:58:20 crc kubenswrapper[4809]: I1124 06:58:20.571241 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bnt6m"] Nov 24 06:58:20 crc kubenswrapper[4809]: I1124 06:58:20.575502 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bnt6m"] Nov 24 06:58:20 crc kubenswrapper[4809]: I1124 06:58:20.604128 4809 scope.go:117] "RemoveContainer" containerID="2be25d53b63d196c5855c6d2768759424b8b21f2b5c00ec3ed4aa14a2da35c4a" Nov 24 06:58:20 crc kubenswrapper[4809]: I1124 06:58:20.900606 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb64a807-42bd-4297-aa31-52a142255d89" path="/var/lib/kubelet/pods/fb64a807-42bd-4297-aa31-52a142255d89/volumes" Nov 24 06:58:21 crc kubenswrapper[4809]: I1124 06:58:21.531560 4809 generic.go:334] "Generic (PLEG): container finished" podID="9c707d1d-d2c2-42db-82b9-afe8fde54914" containerID="cae761370ec4076c7c309bb9c17191b404ce1365f20d1659b4595908a484cd10" exitCode=0 Nov 24 06:58:21 crc kubenswrapper[4809]: I1124 06:58:21.531631 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9r2kr" event={"ID":"9c707d1d-d2c2-42db-82b9-afe8fde54914","Type":"ContainerDied","Data":"cae761370ec4076c7c309bb9c17191b404ce1365f20d1659b4595908a484cd10"} Nov 24 06:58:21 crc kubenswrapper[4809]: I1124 06:58:21.533922 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7rrb4" event={"ID":"c385307d-febf-4fe4-bdf6-51224328ea10","Type":"ContainerStarted","Data":"e1a585654303624241ea74e3c9a6b9d8ef8a9abf5db2f54eaf37cb82a1f5e36f"} Nov 24 06:58:21 crc kubenswrapper[4809]: I1124 06:58:21.537412 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsc5l" event={"ID":"ae7aaa0b-3da0-4764-808a-619dae2fe808","Type":"ContainerStarted","Data":"11d2beca7140e69ba1bd7817a34ca8e127d922b36d5605c08aa05deb627baea0"} Nov 24 06:58:21 crc kubenswrapper[4809]: I1124 06:58:21.540243 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9ttsc" event={"ID":"8c69d61e-4b0d-4388-aff9-9571d2a9c16c","Type":"ContainerStarted","Data":"8843cc17e463173c7aabf2b0c84fac0b5f52425a892b7c0bdd47f092292233bd"} Nov 24 06:58:21 crc kubenswrapper[4809]: I1124 06:58:21.541790 4809 generic.go:334] "Generic (PLEG): container finished" podID="f8951296-b323-4ac3-9e01-963b60ff998c" containerID="abeacc8c9f08cd94466ce62e0c61f64d136cc0e42176bff9132abc9f7ef03c83" exitCode=0 Nov 24 06:58:21 crc kubenswrapper[4809]: I1124 06:58:21.541819 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llxlp" event={"ID":"f8951296-b323-4ac3-9e01-963b60ff998c","Type":"ContainerDied","Data":"abeacc8c9f08cd94466ce62e0c61f64d136cc0e42176bff9132abc9f7ef03c83"} Nov 24 06:58:21 crc kubenswrapper[4809]: I1124 06:58:21.571936 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wsc5l" podStartSLOduration=1.7358495010000001 podStartE2EDuration="50.571897111s" podCreationTimestamp="2025-11-24 06:57:31 +0000 UTC" firstStartedPulling="2025-11-24 06:57:32.1335259 +0000 UTC m=+152.034117605" lastFinishedPulling="2025-11-24 06:58:20.9695735 +0000 UTC m=+200.870165215" observedRunningTime="2025-11-24 06:58:21.566837501 +0000 UTC m=+201.467429226" watchObservedRunningTime="2025-11-24 06:58:21.571897111 +0000 UTC m=+201.472488826" Nov 24 06:58:21 crc kubenswrapper[4809]: I1124 06:58:21.610452 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9ttsc" podStartSLOduration=2.670784403 podStartE2EDuration="53.610433881s" podCreationTimestamp="2025-11-24 06:57:28 +0000 UTC" firstStartedPulling="2025-11-24 06:57:30.043333972 +0000 UTC m=+149.943925677" lastFinishedPulling="2025-11-24 06:58:20.98298345 +0000 UTC m=+200.883575155" observedRunningTime="2025-11-24 06:58:21.60738381 +0000 UTC m=+201.507975515" watchObservedRunningTime="2025-11-24 06:58:21.610433881 +0000 UTC m=+201.511025586" Nov 24 06:58:21 crc kubenswrapper[4809]: I1124 06:58:21.623856 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7rrb4" podStartSLOduration=1.733129664 podStartE2EDuration="51.623840041s" podCreationTimestamp="2025-11-24 06:57:30 +0000 UTC" firstStartedPulling="2025-11-24 06:57:31.094916482 +0000 UTC m=+150.995508187" lastFinishedPulling="2025-11-24 06:58:20.985626849 +0000 UTC m=+200.886218564" observedRunningTime="2025-11-24 06:58:21.622643935 +0000 UTC m=+201.523235640" watchObservedRunningTime="2025-11-24 06:58:21.623840041 +0000 UTC m=+201.524431746" Nov 24 06:58:21 crc kubenswrapper[4809]: I1124 06:58:21.840114 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kppbx" Nov 24 06:58:21 crc kubenswrapper[4809]: I1124 06:58:21.923237 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kppbx" Nov 24 06:58:22 crc kubenswrapper[4809]: I1124 06:58:22.560341 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9r2kr" event={"ID":"9c707d1d-d2c2-42db-82b9-afe8fde54914","Type":"ContainerStarted","Data":"1b606756ec79d1f682e5df4032fefde8a8a0aac4d23d4446bf79e9e28d8d07a1"} Nov 24 06:58:22 crc kubenswrapper[4809]: I1124 06:58:22.562241 4809 generic.go:334] "Generic (PLEG): container finished" podID="1383ed81-5c2a-4416-8c67-05e4881db7be" containerID="b46af9efba74659737669b84699e3efa267ab313b341e0e4fd52a019312e2221" exitCode=0 Nov 24 06:58:22 crc kubenswrapper[4809]: I1124 06:58:22.562320 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pj46g" event={"ID":"1383ed81-5c2a-4416-8c67-05e4881db7be","Type":"ContainerDied","Data":"b46af9efba74659737669b84699e3efa267ab313b341e0e4fd52a019312e2221"} Nov 24 06:58:22 crc kubenswrapper[4809]: I1124 06:58:22.564828 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llxlp" event={"ID":"f8951296-b323-4ac3-9e01-963b60ff998c","Type":"ContainerStarted","Data":"05d73028a172bc3b0d3cd3d56d5f7ca20031d690d99abc7f0ed30e121e7592f4"} Nov 24 06:58:22 crc kubenswrapper[4809]: I1124 06:58:22.583667 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9r2kr" podStartSLOduration=2.756059803 podStartE2EDuration="53.583649748s" podCreationTimestamp="2025-11-24 06:57:29 +0000 UTC" firstStartedPulling="2025-11-24 06:57:31.081692943 +0000 UTC m=+150.982284638" lastFinishedPulling="2025-11-24 06:58:21.909282878 +0000 UTC m=+201.809874583" observedRunningTime="2025-11-24 06:58:22.580388951 +0000 UTC m=+202.480980656" watchObservedRunningTime="2025-11-24 06:58:22.583649748 +0000 UTC m=+202.484241453" Nov 24 06:58:22 crc kubenswrapper[4809]: I1124 06:58:22.601914 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-llxlp" podStartSLOduration=2.649288992 podStartE2EDuration="55.601896623s" podCreationTimestamp="2025-11-24 06:57:27 +0000 UTC" firstStartedPulling="2025-11-24 06:57:29.013739399 +0000 UTC m=+148.914331104" lastFinishedPulling="2025-11-24 06:58:21.96634703 +0000 UTC m=+201.866938735" observedRunningTime="2025-11-24 06:58:22.600273334 +0000 UTC m=+202.500865039" watchObservedRunningTime="2025-11-24 06:58:22.601896623 +0000 UTC m=+202.502488328" Nov 24 06:58:23 crc kubenswrapper[4809]: I1124 06:58:23.571808 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pj46g" event={"ID":"1383ed81-5c2a-4416-8c67-05e4881db7be","Type":"ContainerStarted","Data":"f69124e68a7c6e2ee6c394c6f522c6b844a3e301c21c65411b5d3be131f02e34"} Nov 24 06:58:23 crc kubenswrapper[4809]: I1124 06:58:23.598805 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pj46g" podStartSLOduration=2.188756308 podStartE2EDuration="56.598785627s" podCreationTimestamp="2025-11-24 06:57:27 +0000 UTC" firstStartedPulling="2025-11-24 06:57:29.01991423 +0000 UTC m=+148.920505925" lastFinishedPulling="2025-11-24 06:58:23.429943539 +0000 UTC m=+203.330535244" observedRunningTime="2025-11-24 06:58:23.596473998 +0000 UTC m=+203.497065703" watchObservedRunningTime="2025-11-24 06:58:23.598785627 +0000 UTC m=+203.499377332" Nov 24 06:58:23 crc kubenswrapper[4809]: I1124 06:58:23.721388 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kppbx"] Nov 24 06:58:23 crc kubenswrapper[4809]: I1124 06:58:23.721697 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kppbx" podUID="752560ef-f411-46b8-b3e7-d061cdc1e68b" containerName="registry-server" containerID="cri-o://3c63c013b1d02bca0030cc98821d18401b6f6106e814927ac3adfe3c7a0e09c3" gracePeriod=2 Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.076812 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kppbx" Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.204427 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/752560ef-f411-46b8-b3e7-d061cdc1e68b-catalog-content\") pod \"752560ef-f411-46b8-b3e7-d061cdc1e68b\" (UID: \"752560ef-f411-46b8-b3e7-d061cdc1e68b\") " Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.204510 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/752560ef-f411-46b8-b3e7-d061cdc1e68b-utilities\") pod \"752560ef-f411-46b8-b3e7-d061cdc1e68b\" (UID: \"752560ef-f411-46b8-b3e7-d061cdc1e68b\") " Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.204565 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jv2vk\" (UniqueName: \"kubernetes.io/projected/752560ef-f411-46b8-b3e7-d061cdc1e68b-kube-api-access-jv2vk\") pod \"752560ef-f411-46b8-b3e7-d061cdc1e68b\" (UID: \"752560ef-f411-46b8-b3e7-d061cdc1e68b\") " Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.205144 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/752560ef-f411-46b8-b3e7-d061cdc1e68b-utilities" (OuterVolumeSpecName: "utilities") pod "752560ef-f411-46b8-b3e7-d061cdc1e68b" (UID: "752560ef-f411-46b8-b3e7-d061cdc1e68b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.213089 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/752560ef-f411-46b8-b3e7-d061cdc1e68b-kube-api-access-jv2vk" (OuterVolumeSpecName: "kube-api-access-jv2vk") pod "752560ef-f411-46b8-b3e7-d061cdc1e68b" (UID: "752560ef-f411-46b8-b3e7-d061cdc1e68b"). InnerVolumeSpecName "kube-api-access-jv2vk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.296349 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/752560ef-f411-46b8-b3e7-d061cdc1e68b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "752560ef-f411-46b8-b3e7-d061cdc1e68b" (UID: "752560ef-f411-46b8-b3e7-d061cdc1e68b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.305950 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jv2vk\" (UniqueName: \"kubernetes.io/projected/752560ef-f411-46b8-b3e7-d061cdc1e68b-kube-api-access-jv2vk\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.306019 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/752560ef-f411-46b8-b3e7-d061cdc1e68b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.306033 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/752560ef-f411-46b8-b3e7-d061cdc1e68b-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.578770 4809 generic.go:334] "Generic (PLEG): container finished" podID="752560ef-f411-46b8-b3e7-d061cdc1e68b" containerID="3c63c013b1d02bca0030cc98821d18401b6f6106e814927ac3adfe3c7a0e09c3" exitCode=0 Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.578809 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kppbx" event={"ID":"752560ef-f411-46b8-b3e7-d061cdc1e68b","Type":"ContainerDied","Data":"3c63c013b1d02bca0030cc98821d18401b6f6106e814927ac3adfe3c7a0e09c3"} Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.578836 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kppbx" event={"ID":"752560ef-f411-46b8-b3e7-d061cdc1e68b","Type":"ContainerDied","Data":"fedb1888f481d996552a6c9ff34fddd0eab8dbaeb5aaa1b825fb391604876884"} Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.578851 4809 scope.go:117] "RemoveContainer" containerID="3c63c013b1d02bca0030cc98821d18401b6f6106e814927ac3adfe3c7a0e09c3" Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.578947 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kppbx" Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.599663 4809 scope.go:117] "RemoveContainer" containerID="ec44bd017adf72220beb0ffb97c95012466c69749ff09d0b8b1d316bf1a554a3" Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.604068 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kppbx"] Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.609660 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kppbx"] Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.625056 4809 scope.go:117] "RemoveContainer" containerID="a44d1727650fb8334c7dc28b2da024870ca16c10ea19c90c4c131989a11386f7" Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.640565 4809 scope.go:117] "RemoveContainer" containerID="3c63c013b1d02bca0030cc98821d18401b6f6106e814927ac3adfe3c7a0e09c3" Nov 24 06:58:24 crc kubenswrapper[4809]: E1124 06:58:24.641275 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c63c013b1d02bca0030cc98821d18401b6f6106e814927ac3adfe3c7a0e09c3\": container with ID starting with 3c63c013b1d02bca0030cc98821d18401b6f6106e814927ac3adfe3c7a0e09c3 not found: ID does not exist" containerID="3c63c013b1d02bca0030cc98821d18401b6f6106e814927ac3adfe3c7a0e09c3" Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.641332 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c63c013b1d02bca0030cc98821d18401b6f6106e814927ac3adfe3c7a0e09c3"} err="failed to get container status \"3c63c013b1d02bca0030cc98821d18401b6f6106e814927ac3adfe3c7a0e09c3\": rpc error: code = NotFound desc = could not find container \"3c63c013b1d02bca0030cc98821d18401b6f6106e814927ac3adfe3c7a0e09c3\": container with ID starting with 3c63c013b1d02bca0030cc98821d18401b6f6106e814927ac3adfe3c7a0e09c3 not found: ID does not exist" Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.641479 4809 scope.go:117] "RemoveContainer" containerID="ec44bd017adf72220beb0ffb97c95012466c69749ff09d0b8b1d316bf1a554a3" Nov 24 06:58:24 crc kubenswrapper[4809]: E1124 06:58:24.641871 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec44bd017adf72220beb0ffb97c95012466c69749ff09d0b8b1d316bf1a554a3\": container with ID starting with ec44bd017adf72220beb0ffb97c95012466c69749ff09d0b8b1d316bf1a554a3 not found: ID does not exist" containerID="ec44bd017adf72220beb0ffb97c95012466c69749ff09d0b8b1d316bf1a554a3" Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.641903 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec44bd017adf72220beb0ffb97c95012466c69749ff09d0b8b1d316bf1a554a3"} err="failed to get container status \"ec44bd017adf72220beb0ffb97c95012466c69749ff09d0b8b1d316bf1a554a3\": rpc error: code = NotFound desc = could not find container \"ec44bd017adf72220beb0ffb97c95012466c69749ff09d0b8b1d316bf1a554a3\": container with ID starting with ec44bd017adf72220beb0ffb97c95012466c69749ff09d0b8b1d316bf1a554a3 not found: ID does not exist" Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.641924 4809 scope.go:117] "RemoveContainer" containerID="a44d1727650fb8334c7dc28b2da024870ca16c10ea19c90c4c131989a11386f7" Nov 24 06:58:24 crc kubenswrapper[4809]: E1124 06:58:24.642282 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a44d1727650fb8334c7dc28b2da024870ca16c10ea19c90c4c131989a11386f7\": container with ID starting with a44d1727650fb8334c7dc28b2da024870ca16c10ea19c90c4c131989a11386f7 not found: ID does not exist" containerID="a44d1727650fb8334c7dc28b2da024870ca16c10ea19c90c4c131989a11386f7" Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.642320 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a44d1727650fb8334c7dc28b2da024870ca16c10ea19c90c4c131989a11386f7"} err="failed to get container status \"a44d1727650fb8334c7dc28b2da024870ca16c10ea19c90c4c131989a11386f7\": rpc error: code = NotFound desc = could not find container \"a44d1727650fb8334c7dc28b2da024870ca16c10ea19c90c4c131989a11386f7\": container with ID starting with a44d1727650fb8334c7dc28b2da024870ca16c10ea19c90c4c131989a11386f7 not found: ID does not exist" Nov 24 06:58:24 crc kubenswrapper[4809]: I1124 06:58:24.901719 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="752560ef-f411-46b8-b3e7-d061cdc1e68b" path="/var/lib/kubelet/pods/752560ef-f411-46b8-b3e7-d061cdc1e68b/volumes" Nov 24 06:58:27 crc kubenswrapper[4809]: I1124 06:58:27.966915 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pj46g" Nov 24 06:58:27 crc kubenswrapper[4809]: I1124 06:58:27.967277 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pj46g" Nov 24 06:58:28 crc kubenswrapper[4809]: I1124 06:58:28.011670 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pj46g" Nov 24 06:58:28 crc kubenswrapper[4809]: I1124 06:58:28.189157 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-llxlp" Nov 24 06:58:28 crc kubenswrapper[4809]: I1124 06:58:28.189477 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-llxlp" Nov 24 06:58:28 crc kubenswrapper[4809]: I1124 06:58:28.229796 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-llxlp" Nov 24 06:58:28 crc kubenswrapper[4809]: I1124 06:58:28.588038 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9ttsc" Nov 24 06:58:28 crc kubenswrapper[4809]: I1124 06:58:28.588086 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9ttsc" Nov 24 06:58:28 crc kubenswrapper[4809]: I1124 06:58:28.640502 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9ttsc" Nov 24 06:58:28 crc kubenswrapper[4809]: I1124 06:58:28.654296 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pj46g" Nov 24 06:58:28 crc kubenswrapper[4809]: I1124 06:58:28.666584 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-llxlp" Nov 24 06:58:28 crc kubenswrapper[4809]: I1124 06:58:28.686945 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9ttsc" Nov 24 06:58:29 crc kubenswrapper[4809]: I1124 06:58:29.952862 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9r2kr" Nov 24 06:58:29 crc kubenswrapper[4809]: I1124 06:58:29.952936 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9r2kr" Nov 24 06:58:30 crc kubenswrapper[4809]: I1124 06:58:30.000837 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9r2kr" Nov 24 06:58:30 crc kubenswrapper[4809]: I1124 06:58:30.415415 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7rrb4" Nov 24 06:58:30 crc kubenswrapper[4809]: I1124 06:58:30.415470 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7rrb4" Nov 24 06:58:30 crc kubenswrapper[4809]: I1124 06:58:30.468904 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7rrb4" Nov 24 06:58:30 crc kubenswrapper[4809]: I1124 06:58:30.518729 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9ttsc"] Nov 24 06:58:30 crc kubenswrapper[4809]: I1124 06:58:30.626227 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9ttsc" podUID="8c69d61e-4b0d-4388-aff9-9571d2a9c16c" containerName="registry-server" containerID="cri-o://8843cc17e463173c7aabf2b0c84fac0b5f52425a892b7c0bdd47f092292233bd" gracePeriod=2 Nov 24 06:58:30 crc kubenswrapper[4809]: I1124 06:58:30.672241 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9r2kr" Nov 24 06:58:30 crc kubenswrapper[4809]: I1124 06:58:30.683839 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7rrb4" Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.000086 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9ttsc" Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.088235 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c69d61e-4b0d-4388-aff9-9571d2a9c16c-catalog-content\") pod \"8c69d61e-4b0d-4388-aff9-9571d2a9c16c\" (UID: \"8c69d61e-4b0d-4388-aff9-9571d2a9c16c\") " Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.088324 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vm75s\" (UniqueName: \"kubernetes.io/projected/8c69d61e-4b0d-4388-aff9-9571d2a9c16c-kube-api-access-vm75s\") pod \"8c69d61e-4b0d-4388-aff9-9571d2a9c16c\" (UID: \"8c69d61e-4b0d-4388-aff9-9571d2a9c16c\") " Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.088435 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c69d61e-4b0d-4388-aff9-9571d2a9c16c-utilities\") pod \"8c69d61e-4b0d-4388-aff9-9571d2a9c16c\" (UID: \"8c69d61e-4b0d-4388-aff9-9571d2a9c16c\") " Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.089345 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c69d61e-4b0d-4388-aff9-9571d2a9c16c-utilities" (OuterVolumeSpecName: "utilities") pod "8c69d61e-4b0d-4388-aff9-9571d2a9c16c" (UID: "8c69d61e-4b0d-4388-aff9-9571d2a9c16c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.093096 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c69d61e-4b0d-4388-aff9-9571d2a9c16c-kube-api-access-vm75s" (OuterVolumeSpecName: "kube-api-access-vm75s") pod "8c69d61e-4b0d-4388-aff9-9571d2a9c16c" (UID: "8c69d61e-4b0d-4388-aff9-9571d2a9c16c"). InnerVolumeSpecName "kube-api-access-vm75s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.141473 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c69d61e-4b0d-4388-aff9-9571d2a9c16c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c69d61e-4b0d-4388-aff9-9571d2a9c16c" (UID: "8c69d61e-4b0d-4388-aff9-9571d2a9c16c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.189516 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c69d61e-4b0d-4388-aff9-9571d2a9c16c-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.189551 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c69d61e-4b0d-4388-aff9-9571d2a9c16c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.189565 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vm75s\" (UniqueName: \"kubernetes.io/projected/8c69d61e-4b0d-4388-aff9-9571d2a9c16c-kube-api-access-vm75s\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.382397 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wsc5l" Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.382430 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wsc5l" Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.444337 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wsc5l" Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.633322 4809 generic.go:334] "Generic (PLEG): container finished" podID="8c69d61e-4b0d-4388-aff9-9571d2a9c16c" containerID="8843cc17e463173c7aabf2b0c84fac0b5f52425a892b7c0bdd47f092292233bd" exitCode=0 Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.633390 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9ttsc" Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.633490 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9ttsc" event={"ID":"8c69d61e-4b0d-4388-aff9-9571d2a9c16c","Type":"ContainerDied","Data":"8843cc17e463173c7aabf2b0c84fac0b5f52425a892b7c0bdd47f092292233bd"} Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.633538 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9ttsc" event={"ID":"8c69d61e-4b0d-4388-aff9-9571d2a9c16c","Type":"ContainerDied","Data":"31f65ad8a0d43722450c775e01cc7a8f016e1f8c04bc5ab6b856b179b7d7da1c"} Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.633560 4809 scope.go:117] "RemoveContainer" containerID="8843cc17e463173c7aabf2b0c84fac0b5f52425a892b7c0bdd47f092292233bd" Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.662295 4809 scope.go:117] "RemoveContainer" containerID="c3c577bbcad29f4843a716708cc19b98cba12bbbc5ae59a46de75fd96c30b961" Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.668778 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9ttsc"] Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.671188 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9ttsc"] Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.674025 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wsc5l" Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.704566 4809 scope.go:117] "RemoveContainer" containerID="d00b465a0808d4b37f877b29f106cf72c4d30f0410672f93e34a2c3d6c1c2912" Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.744828 4809 scope.go:117] "RemoveContainer" containerID="8843cc17e463173c7aabf2b0c84fac0b5f52425a892b7c0bdd47f092292233bd" Nov 24 06:58:31 crc kubenswrapper[4809]: E1124 06:58:31.748567 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8843cc17e463173c7aabf2b0c84fac0b5f52425a892b7c0bdd47f092292233bd\": container with ID starting with 8843cc17e463173c7aabf2b0c84fac0b5f52425a892b7c0bdd47f092292233bd not found: ID does not exist" containerID="8843cc17e463173c7aabf2b0c84fac0b5f52425a892b7c0bdd47f092292233bd" Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.748619 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8843cc17e463173c7aabf2b0c84fac0b5f52425a892b7c0bdd47f092292233bd"} err="failed to get container status \"8843cc17e463173c7aabf2b0c84fac0b5f52425a892b7c0bdd47f092292233bd\": rpc error: code = NotFound desc = could not find container \"8843cc17e463173c7aabf2b0c84fac0b5f52425a892b7c0bdd47f092292233bd\": container with ID starting with 8843cc17e463173c7aabf2b0c84fac0b5f52425a892b7c0bdd47f092292233bd not found: ID does not exist" Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.748648 4809 scope.go:117] "RemoveContainer" containerID="c3c577bbcad29f4843a716708cc19b98cba12bbbc5ae59a46de75fd96c30b961" Nov 24 06:58:31 crc kubenswrapper[4809]: E1124 06:58:31.756190 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3c577bbcad29f4843a716708cc19b98cba12bbbc5ae59a46de75fd96c30b961\": container with ID starting with c3c577bbcad29f4843a716708cc19b98cba12bbbc5ae59a46de75fd96c30b961 not found: ID does not exist" containerID="c3c577bbcad29f4843a716708cc19b98cba12bbbc5ae59a46de75fd96c30b961" Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.756251 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3c577bbcad29f4843a716708cc19b98cba12bbbc5ae59a46de75fd96c30b961"} err="failed to get container status \"c3c577bbcad29f4843a716708cc19b98cba12bbbc5ae59a46de75fd96c30b961\": rpc error: code = NotFound desc = could not find container \"c3c577bbcad29f4843a716708cc19b98cba12bbbc5ae59a46de75fd96c30b961\": container with ID starting with c3c577bbcad29f4843a716708cc19b98cba12bbbc5ae59a46de75fd96c30b961 not found: ID does not exist" Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.756285 4809 scope.go:117] "RemoveContainer" containerID="d00b465a0808d4b37f877b29f106cf72c4d30f0410672f93e34a2c3d6c1c2912" Nov 24 06:58:31 crc kubenswrapper[4809]: E1124 06:58:31.760501 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d00b465a0808d4b37f877b29f106cf72c4d30f0410672f93e34a2c3d6c1c2912\": container with ID starting with d00b465a0808d4b37f877b29f106cf72c4d30f0410672f93e34a2c3d6c1c2912 not found: ID does not exist" containerID="d00b465a0808d4b37f877b29f106cf72c4d30f0410672f93e34a2c3d6c1c2912" Nov 24 06:58:31 crc kubenswrapper[4809]: I1124 06:58:31.760547 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d00b465a0808d4b37f877b29f106cf72c4d30f0410672f93e34a2c3d6c1c2912"} err="failed to get container status \"d00b465a0808d4b37f877b29f106cf72c4d30f0410672f93e34a2c3d6c1c2912\": rpc error: code = NotFound desc = could not find container \"d00b465a0808d4b37f877b29f106cf72c4d30f0410672f93e34a2c3d6c1c2912\": container with ID starting with d00b465a0808d4b37f877b29f106cf72c4d30f0410672f93e34a2c3d6c1c2912 not found: ID does not exist" Nov 24 06:58:32 crc kubenswrapper[4809]: I1124 06:58:32.722058 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7rrb4"] Nov 24 06:58:32 crc kubenswrapper[4809]: I1124 06:58:32.722573 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7rrb4" podUID="c385307d-febf-4fe4-bdf6-51224328ea10" containerName="registry-server" containerID="cri-o://e1a585654303624241ea74e3c9a6b9d8ef8a9abf5db2f54eaf37cb82a1f5e36f" gracePeriod=2 Nov 24 06:58:32 crc kubenswrapper[4809]: I1124 06:58:32.897342 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c69d61e-4b0d-4388-aff9-9571d2a9c16c" path="/var/lib/kubelet/pods/8c69d61e-4b0d-4388-aff9-9571d2a9c16c/volumes" Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.067038 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7rrb4" Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.118501 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c385307d-febf-4fe4-bdf6-51224328ea10-utilities\") pod \"c385307d-febf-4fe4-bdf6-51224328ea10\" (UID: \"c385307d-febf-4fe4-bdf6-51224328ea10\") " Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.118613 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcbrr\" (UniqueName: \"kubernetes.io/projected/c385307d-febf-4fe4-bdf6-51224328ea10-kube-api-access-kcbrr\") pod \"c385307d-febf-4fe4-bdf6-51224328ea10\" (UID: \"c385307d-febf-4fe4-bdf6-51224328ea10\") " Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.118635 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c385307d-febf-4fe4-bdf6-51224328ea10-catalog-content\") pod \"c385307d-febf-4fe4-bdf6-51224328ea10\" (UID: \"c385307d-febf-4fe4-bdf6-51224328ea10\") " Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.119222 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c385307d-febf-4fe4-bdf6-51224328ea10-utilities" (OuterVolumeSpecName: "utilities") pod "c385307d-febf-4fe4-bdf6-51224328ea10" (UID: "c385307d-febf-4fe4-bdf6-51224328ea10"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.123139 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c385307d-febf-4fe4-bdf6-51224328ea10-kube-api-access-kcbrr" (OuterVolumeSpecName: "kube-api-access-kcbrr") pod "c385307d-febf-4fe4-bdf6-51224328ea10" (UID: "c385307d-febf-4fe4-bdf6-51224328ea10"). InnerVolumeSpecName "kube-api-access-kcbrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.135094 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c385307d-febf-4fe4-bdf6-51224328ea10-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c385307d-febf-4fe4-bdf6-51224328ea10" (UID: "c385307d-febf-4fe4-bdf6-51224328ea10"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.219530 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c385307d-febf-4fe4-bdf6-51224328ea10-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.219574 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c385307d-febf-4fe4-bdf6-51224328ea10-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.219587 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcbrr\" (UniqueName: \"kubernetes.io/projected/c385307d-febf-4fe4-bdf6-51224328ea10-kube-api-access-kcbrr\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.651631 4809 generic.go:334] "Generic (PLEG): container finished" podID="c385307d-febf-4fe4-bdf6-51224328ea10" containerID="e1a585654303624241ea74e3c9a6b9d8ef8a9abf5db2f54eaf37cb82a1f5e36f" exitCode=0 Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.651731 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7rrb4" event={"ID":"c385307d-febf-4fe4-bdf6-51224328ea10","Type":"ContainerDied","Data":"e1a585654303624241ea74e3c9a6b9d8ef8a9abf5db2f54eaf37cb82a1f5e36f"} Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.651950 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7rrb4" event={"ID":"c385307d-febf-4fe4-bdf6-51224328ea10","Type":"ContainerDied","Data":"0858b5e27e106a15a0a62bbb489b55fc0200aecbc20568a1f6461bbec7852d1c"} Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.651769 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7rrb4" Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.651991 4809 scope.go:117] "RemoveContainer" containerID="e1a585654303624241ea74e3c9a6b9d8ef8a9abf5db2f54eaf37cb82a1f5e36f" Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.669911 4809 scope.go:117] "RemoveContainer" containerID="5c288b1bcd9d858a0d768a8816cee9c7a4ee05bc910ed2b73214a7af339b4ba9" Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.686336 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7rrb4"] Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.693051 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7rrb4"] Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.693217 4809 scope.go:117] "RemoveContainer" containerID="19e5a66331aea317b795426c2fd47980fea970ffa24b10413d5d74c4ae8acf9b" Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.726838 4809 scope.go:117] "RemoveContainer" containerID="e1a585654303624241ea74e3c9a6b9d8ef8a9abf5db2f54eaf37cb82a1f5e36f" Nov 24 06:58:33 crc kubenswrapper[4809]: E1124 06:58:33.727443 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1a585654303624241ea74e3c9a6b9d8ef8a9abf5db2f54eaf37cb82a1f5e36f\": container with ID starting with e1a585654303624241ea74e3c9a6b9d8ef8a9abf5db2f54eaf37cb82a1f5e36f not found: ID does not exist" containerID="e1a585654303624241ea74e3c9a6b9d8ef8a9abf5db2f54eaf37cb82a1f5e36f" Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.727472 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1a585654303624241ea74e3c9a6b9d8ef8a9abf5db2f54eaf37cb82a1f5e36f"} err="failed to get container status \"e1a585654303624241ea74e3c9a6b9d8ef8a9abf5db2f54eaf37cb82a1f5e36f\": rpc error: code = NotFound desc = could not find container \"e1a585654303624241ea74e3c9a6b9d8ef8a9abf5db2f54eaf37cb82a1f5e36f\": container with ID starting with e1a585654303624241ea74e3c9a6b9d8ef8a9abf5db2f54eaf37cb82a1f5e36f not found: ID does not exist" Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.727498 4809 scope.go:117] "RemoveContainer" containerID="5c288b1bcd9d858a0d768a8816cee9c7a4ee05bc910ed2b73214a7af339b4ba9" Nov 24 06:58:33 crc kubenswrapper[4809]: E1124 06:58:33.728058 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c288b1bcd9d858a0d768a8816cee9c7a4ee05bc910ed2b73214a7af339b4ba9\": container with ID starting with 5c288b1bcd9d858a0d768a8816cee9c7a4ee05bc910ed2b73214a7af339b4ba9 not found: ID does not exist" containerID="5c288b1bcd9d858a0d768a8816cee9c7a4ee05bc910ed2b73214a7af339b4ba9" Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.728096 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c288b1bcd9d858a0d768a8816cee9c7a4ee05bc910ed2b73214a7af339b4ba9"} err="failed to get container status \"5c288b1bcd9d858a0d768a8816cee9c7a4ee05bc910ed2b73214a7af339b4ba9\": rpc error: code = NotFound desc = could not find container \"5c288b1bcd9d858a0d768a8816cee9c7a4ee05bc910ed2b73214a7af339b4ba9\": container with ID starting with 5c288b1bcd9d858a0d768a8816cee9c7a4ee05bc910ed2b73214a7af339b4ba9 not found: ID does not exist" Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.728123 4809 scope.go:117] "RemoveContainer" containerID="19e5a66331aea317b795426c2fd47980fea970ffa24b10413d5d74c4ae8acf9b" Nov 24 06:58:33 crc kubenswrapper[4809]: E1124 06:58:33.728479 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19e5a66331aea317b795426c2fd47980fea970ffa24b10413d5d74c4ae8acf9b\": container with ID starting with 19e5a66331aea317b795426c2fd47980fea970ffa24b10413d5d74c4ae8acf9b not found: ID does not exist" containerID="19e5a66331aea317b795426c2fd47980fea970ffa24b10413d5d74c4ae8acf9b" Nov 24 06:58:33 crc kubenswrapper[4809]: I1124 06:58:33.728495 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19e5a66331aea317b795426c2fd47980fea970ffa24b10413d5d74c4ae8acf9b"} err="failed to get container status \"19e5a66331aea317b795426c2fd47980fea970ffa24b10413d5d74c4ae8acf9b\": rpc error: code = NotFound desc = could not find container \"19e5a66331aea317b795426c2fd47980fea970ffa24b10413d5d74c4ae8acf9b\": container with ID starting with 19e5a66331aea317b795426c2fd47980fea970ffa24b10413d5d74c4ae8acf9b not found: ID does not exist" Nov 24 06:58:34 crc kubenswrapper[4809]: I1124 06:58:34.898392 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c385307d-febf-4fe4-bdf6-51224328ea10" path="/var/lib/kubelet/pods/c385307d-febf-4fe4-bdf6-51224328ea10/volumes" Nov 24 06:58:40 crc kubenswrapper[4809]: I1124 06:58:40.830997 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-69dgx"] Nov 24 06:58:48 crc kubenswrapper[4809]: I1124 06:58:48.043683 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 06:58:48 crc kubenswrapper[4809]: I1124 06:58:48.044135 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 06:58:48 crc kubenswrapper[4809]: I1124 06:58:48.044183 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 06:58:48 crc kubenswrapper[4809]: I1124 06:58:48.044718 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5"} pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 06:58:48 crc kubenswrapper[4809]: I1124 06:58:48.044772 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" containerID="cri-o://a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5" gracePeriod=600 Nov 24 06:58:48 crc kubenswrapper[4809]: I1124 06:58:48.736264 4809 generic.go:334] "Generic (PLEG): container finished" podID="be8050d5-1aea-453e-a06e-87702b0e856b" containerID="a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5" exitCode=0 Nov 24 06:58:48 crc kubenswrapper[4809]: I1124 06:58:48.736356 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerDied","Data":"a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5"} Nov 24 06:58:48 crc kubenswrapper[4809]: I1124 06:58:48.736638 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerStarted","Data":"33316ec1d7739e9081fef97a1f1a3a1c2f00b3bbab2a32058a1c6e7ff1f92aad"} Nov 24 06:59:05 crc kubenswrapper[4809]: I1124 06:59:05.875540 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" podUID="51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf" containerName="oauth-openshift" containerID="cri-o://473a63602d48c8f4db20c57684dd8ed5918e2686427a121d8c5c7f87130c5d2b" gracePeriod=15 Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.269696 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.304308 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-86449fd7db-7btdl"] Nov 24 06:59:06 crc kubenswrapper[4809]: E1124 06:59:06.304587 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c385307d-febf-4fe4-bdf6-51224328ea10" containerName="extract-content" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.304605 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c385307d-febf-4fe4-bdf6-51224328ea10" containerName="extract-content" Nov 24 06:59:06 crc kubenswrapper[4809]: E1124 06:59:06.304621 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c69d61e-4b0d-4388-aff9-9571d2a9c16c" containerName="extract-content" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.304630 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c69d61e-4b0d-4388-aff9-9571d2a9c16c" containerName="extract-content" Nov 24 06:59:06 crc kubenswrapper[4809]: E1124 06:59:06.304644 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c69d61e-4b0d-4388-aff9-9571d2a9c16c" containerName="extract-utilities" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.304652 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c69d61e-4b0d-4388-aff9-9571d2a9c16c" containerName="extract-utilities" Nov 24 06:59:06 crc kubenswrapper[4809]: E1124 06:59:06.304661 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="752560ef-f411-46b8-b3e7-d061cdc1e68b" containerName="extract-content" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.304669 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="752560ef-f411-46b8-b3e7-d061cdc1e68b" containerName="extract-content" Nov 24 06:59:06 crc kubenswrapper[4809]: E1124 06:59:06.304683 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c385307d-febf-4fe4-bdf6-51224328ea10" containerName="registry-server" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.304691 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c385307d-febf-4fe4-bdf6-51224328ea10" containerName="registry-server" Nov 24 06:59:06 crc kubenswrapper[4809]: E1124 06:59:06.304701 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb64a807-42bd-4297-aa31-52a142255d89" containerName="extract-utilities" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.304709 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb64a807-42bd-4297-aa31-52a142255d89" containerName="extract-utilities" Nov 24 06:59:06 crc kubenswrapper[4809]: E1124 06:59:06.304721 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c385307d-febf-4fe4-bdf6-51224328ea10" containerName="extract-utilities" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.304729 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c385307d-febf-4fe4-bdf6-51224328ea10" containerName="extract-utilities" Nov 24 06:59:06 crc kubenswrapper[4809]: E1124 06:59:06.304744 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="752560ef-f411-46b8-b3e7-d061cdc1e68b" containerName="extract-utilities" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.304752 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="752560ef-f411-46b8-b3e7-d061cdc1e68b" containerName="extract-utilities" Nov 24 06:59:06 crc kubenswrapper[4809]: E1124 06:59:06.304761 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="752560ef-f411-46b8-b3e7-d061cdc1e68b" containerName="registry-server" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.304769 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="752560ef-f411-46b8-b3e7-d061cdc1e68b" containerName="registry-server" Nov 24 06:59:06 crc kubenswrapper[4809]: E1124 06:59:06.304779 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c69d61e-4b0d-4388-aff9-9571d2a9c16c" containerName="registry-server" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.304786 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c69d61e-4b0d-4388-aff9-9571d2a9c16c" containerName="registry-server" Nov 24 06:59:06 crc kubenswrapper[4809]: E1124 06:59:06.304825 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa" containerName="pruner" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.304833 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa" containerName="pruner" Nov 24 06:59:06 crc kubenswrapper[4809]: E1124 06:59:06.304843 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb64a807-42bd-4297-aa31-52a142255d89" containerName="extract-content" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.304852 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb64a807-42bd-4297-aa31-52a142255d89" containerName="extract-content" Nov 24 06:59:06 crc kubenswrapper[4809]: E1124 06:59:06.304865 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf" containerName="oauth-openshift" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.304873 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf" containerName="oauth-openshift" Nov 24 06:59:06 crc kubenswrapper[4809]: E1124 06:59:06.304882 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="442d2bfc-3172-4bd8-bfb6-f14668b8c971" containerName="pruner" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.304891 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="442d2bfc-3172-4bd8-bfb6-f14668b8c971" containerName="pruner" Nov 24 06:59:06 crc kubenswrapper[4809]: E1124 06:59:06.304904 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb64a807-42bd-4297-aa31-52a142255d89" containerName="registry-server" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.304915 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb64a807-42bd-4297-aa31-52a142255d89" containerName="registry-server" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.305135 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf" containerName="oauth-openshift" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.305154 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="752560ef-f411-46b8-b3e7-d061cdc1e68b" containerName="registry-server" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.305167 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f0b7e2d-ada0-4191-b78b-fb0dcdd5c7aa" containerName="pruner" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.305178 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb64a807-42bd-4297-aa31-52a142255d89" containerName="registry-server" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.305192 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c69d61e-4b0d-4388-aff9-9571d2a9c16c" containerName="registry-server" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.305201 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="442d2bfc-3172-4bd8-bfb6-f14668b8c971" containerName="pruner" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.305213 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c385307d-febf-4fe4-bdf6-51224328ea10" containerName="registry-server" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.305658 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.322717 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-trusted-ca-bundle\") pod \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.322777 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-router-certs\") pod \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.322808 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-template-provider-selection\") pod \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.322834 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-audit-policies\") pod \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.322861 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-template-login\") pod \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.322886 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-audit-dir\") pod \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.322904 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-service-ca\") pod \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.322937 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-serving-cert\") pod \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.322997 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-idp-0-file-data\") pod \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.323052 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-template-error\") pod \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.323102 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbmcq\" (UniqueName: \"kubernetes.io/projected/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-kube-api-access-rbmcq\") pod \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.323128 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-cliconfig\") pod \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.323151 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-session\") pod \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.323185 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-ocp-branding-template\") pod \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\" (UID: \"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf\") " Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.324530 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf" (UID: "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.325109 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf" (UID: "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.327379 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf" (UID: "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.328756 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf" (UID: "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.329440 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf" (UID: "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.330227 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf" (UID: "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.331039 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf" (UID: "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.331339 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf" (UID: "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.332730 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf" (UID: "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.337298 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf" (UID: "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.338168 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-86449fd7db-7btdl"] Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.338505 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf" (UID: "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.339115 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf" (UID: "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.342233 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf" (UID: "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.342261 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-kube-api-access-rbmcq" (OuterVolumeSpecName: "kube-api-access-rbmcq") pod "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf" (UID: "51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf"). InnerVolumeSpecName "kube-api-access-rbmcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424117 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfrqc\" (UniqueName: \"kubernetes.io/projected/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-kube-api-access-zfrqc\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424166 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-system-serving-cert\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424193 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424276 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-system-session\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424335 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-user-template-login\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424362 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-system-cliconfig\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424388 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-user-template-error\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424418 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424496 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-system-router-certs\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424553 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424598 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-audit-policies\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424626 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-system-service-ca\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424661 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424751 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-audit-dir\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424807 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424824 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424839 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424855 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbmcq\" (UniqueName: \"kubernetes.io/projected/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-kube-api-access-rbmcq\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424870 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424883 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424897 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424911 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424935 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424950 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.424987 4809 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.425000 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.425013 4809 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.425024 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.526811 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-audit-dir\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.526933 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-audit-dir\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.526944 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfrqc\" (UniqueName: \"kubernetes.io/projected/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-kube-api-access-zfrqc\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.527057 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-system-serving-cert\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.527081 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.527751 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-system-session\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.527788 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-user-template-login\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.527827 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-system-cliconfig\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.527881 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-user-template-error\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.527937 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.527986 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-system-router-certs\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.528005 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.528108 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-audit-policies\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.528129 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-system-service-ca\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.528176 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.530050 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.530579 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-audit-policies\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.531213 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-system-cliconfig\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.532141 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.532862 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-system-session\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.533063 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-system-service-ca\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.533267 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-system-serving-cert\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.533444 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.533959 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-user-template-error\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.534266 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-user-template-login\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.534677 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-system-router-certs\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.534994 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.548739 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfrqc\" (UniqueName: \"kubernetes.io/projected/6705cb64-25e6-43ff-99ca-cc32cdc1ea21-kube-api-access-zfrqc\") pod \"oauth-openshift-86449fd7db-7btdl\" (UID: \"6705cb64-25e6-43ff-99ca-cc32cdc1ea21\") " pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.628889 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.841582 4809 generic.go:334] "Generic (PLEG): container finished" podID="51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf" containerID="473a63602d48c8f4db20c57684dd8ed5918e2686427a121d8c5c7f87130c5d2b" exitCode=0 Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.841626 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" event={"ID":"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf","Type":"ContainerDied","Data":"473a63602d48c8f4db20c57684dd8ed5918e2686427a121d8c5c7f87130c5d2b"} Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.841652 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.841674 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-69dgx" event={"ID":"51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf","Type":"ContainerDied","Data":"a3277eb73eeb2de4365bdcb3f77cfed5c6cd44a3052bfc77b835dcfc2e3bd4ed"} Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.841697 4809 scope.go:117] "RemoveContainer" containerID="473a63602d48c8f4db20c57684dd8ed5918e2686427a121d8c5c7f87130c5d2b" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.869193 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-69dgx"] Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.872181 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-69dgx"] Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.873463 4809 scope.go:117] "RemoveContainer" containerID="473a63602d48c8f4db20c57684dd8ed5918e2686427a121d8c5c7f87130c5d2b" Nov 24 06:59:06 crc kubenswrapper[4809]: E1124 06:59:06.873940 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"473a63602d48c8f4db20c57684dd8ed5918e2686427a121d8c5c7f87130c5d2b\": container with ID starting with 473a63602d48c8f4db20c57684dd8ed5918e2686427a121d8c5c7f87130c5d2b not found: ID does not exist" containerID="473a63602d48c8f4db20c57684dd8ed5918e2686427a121d8c5c7f87130c5d2b" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.873988 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"473a63602d48c8f4db20c57684dd8ed5918e2686427a121d8c5c7f87130c5d2b"} err="failed to get container status \"473a63602d48c8f4db20c57684dd8ed5918e2686427a121d8c5c7f87130c5d2b\": rpc error: code = NotFound desc = could not find container \"473a63602d48c8f4db20c57684dd8ed5918e2686427a121d8c5c7f87130c5d2b\": container with ID starting with 473a63602d48c8f4db20c57684dd8ed5918e2686427a121d8c5c7f87130c5d2b not found: ID does not exist" Nov 24 06:59:06 crc kubenswrapper[4809]: I1124 06:59:06.898295 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf" path="/var/lib/kubelet/pods/51fd8b7a-774c-4c76-ae45-3fa2f4e7bbaf/volumes" Nov 24 06:59:07 crc kubenswrapper[4809]: I1124 06:59:07.016167 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-86449fd7db-7btdl"] Nov 24 06:59:07 crc kubenswrapper[4809]: I1124 06:59:07.848871 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" event={"ID":"6705cb64-25e6-43ff-99ca-cc32cdc1ea21","Type":"ContainerStarted","Data":"1558a299122323bce8bffaf2987dc18aaf328b0696280af2774a944507d784fd"} Nov 24 06:59:07 crc kubenswrapper[4809]: I1124 06:59:07.849323 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" event={"ID":"6705cb64-25e6-43ff-99ca-cc32cdc1ea21","Type":"ContainerStarted","Data":"91f622638864a9f7c0f20f7730969b8840f33e0ae0bb38f916a1c48d8131f3c7"} Nov 24 06:59:07 crc kubenswrapper[4809]: I1124 06:59:07.850827 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:07 crc kubenswrapper[4809]: I1124 06:59:07.856522 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" Nov 24 06:59:07 crc kubenswrapper[4809]: I1124 06:59:07.873603 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-86449fd7db-7btdl" podStartSLOduration=27.873579059 podStartE2EDuration="27.873579059s" podCreationTimestamp="2025-11-24 06:58:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:59:07.869694594 +0000 UTC m=+247.770286309" watchObservedRunningTime="2025-11-24 06:59:07.873579059 +0000 UTC m=+247.774170794" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.156334 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pj46g"] Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.157132 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pj46g" podUID="1383ed81-5c2a-4416-8c67-05e4881db7be" containerName="registry-server" containerID="cri-o://f69124e68a7c6e2ee6c394c6f522c6b844a3e301c21c65411b5d3be131f02e34" gracePeriod=30 Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.168144 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-llxlp"] Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.168450 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-llxlp" podUID="f8951296-b323-4ac3-9e01-963b60ff998c" containerName="registry-server" containerID="cri-o://05d73028a172bc3b0d3cd3d56d5f7ca20031d690d99abc7f0ed30e121e7592f4" gracePeriod=30 Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.196223 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-svxc5"] Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.196452 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" podUID="b08de060-f5b7-4bed-a27c-89294a8827fa" containerName="marketplace-operator" containerID="cri-o://f5b1bb43ee1cfb566a2b2292ef38df7c63aeaee858c08fee970d588092fb019e" gracePeriod=30 Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.198339 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9r2kr"] Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.198576 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9r2kr" podUID="9c707d1d-d2c2-42db-82b9-afe8fde54914" containerName="registry-server" containerID="cri-o://1b606756ec79d1f682e5df4032fefde8a8a0aac4d23d4446bf79e9e28d8d07a1" gracePeriod=30 Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.206106 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zqwx5"] Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.206711 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wsc5l"] Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.206898 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wsc5l" podUID="ae7aaa0b-3da0-4764-808a-619dae2fe808" containerName="registry-server" containerID="cri-o://11d2beca7140e69ba1bd7817a34ca8e127d922b36d5605c08aa05deb627baea0" gracePeriod=30 Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.207049 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-zqwx5" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.209112 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zqwx5"] Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.282069 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4b7632a5-d284-4a30-89dc-c301a7d13609-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-zqwx5\" (UID: \"4b7632a5-d284-4a30-89dc-c301a7d13609\") " pod="openshift-marketplace/marketplace-operator-79b997595-zqwx5" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.282110 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w262\" (UniqueName: \"kubernetes.io/projected/4b7632a5-d284-4a30-89dc-c301a7d13609-kube-api-access-6w262\") pod \"marketplace-operator-79b997595-zqwx5\" (UID: \"4b7632a5-d284-4a30-89dc-c301a7d13609\") " pod="openshift-marketplace/marketplace-operator-79b997595-zqwx5" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.282249 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4b7632a5-d284-4a30-89dc-c301a7d13609-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-zqwx5\" (UID: \"4b7632a5-d284-4a30-89dc-c301a7d13609\") " pod="openshift-marketplace/marketplace-operator-79b997595-zqwx5" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.382670 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4b7632a5-d284-4a30-89dc-c301a7d13609-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-zqwx5\" (UID: \"4b7632a5-d284-4a30-89dc-c301a7d13609\") " pod="openshift-marketplace/marketplace-operator-79b997595-zqwx5" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.382715 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4b7632a5-d284-4a30-89dc-c301a7d13609-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-zqwx5\" (UID: \"4b7632a5-d284-4a30-89dc-c301a7d13609\") " pod="openshift-marketplace/marketplace-operator-79b997595-zqwx5" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.382760 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w262\" (UniqueName: \"kubernetes.io/projected/4b7632a5-d284-4a30-89dc-c301a7d13609-kube-api-access-6w262\") pod \"marketplace-operator-79b997595-zqwx5\" (UID: \"4b7632a5-d284-4a30-89dc-c301a7d13609\") " pod="openshift-marketplace/marketplace-operator-79b997595-zqwx5" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.384000 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4b7632a5-d284-4a30-89dc-c301a7d13609-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-zqwx5\" (UID: \"4b7632a5-d284-4a30-89dc-c301a7d13609\") " pod="openshift-marketplace/marketplace-operator-79b997595-zqwx5" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.390281 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4b7632a5-d284-4a30-89dc-c301a7d13609-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-zqwx5\" (UID: \"4b7632a5-d284-4a30-89dc-c301a7d13609\") " pod="openshift-marketplace/marketplace-operator-79b997595-zqwx5" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.399227 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w262\" (UniqueName: \"kubernetes.io/projected/4b7632a5-d284-4a30-89dc-c301a7d13609-kube-api-access-6w262\") pod \"marketplace-operator-79b997595-zqwx5\" (UID: \"4b7632a5-d284-4a30-89dc-c301a7d13609\") " pod="openshift-marketplace/marketplace-operator-79b997595-zqwx5" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.590220 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-zqwx5" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.593495 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pj46g" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.644556 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-llxlp" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.651937 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9r2kr" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.656610 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.673423 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wsc5l" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.687859 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae7aaa0b-3da0-4764-808a-619dae2fe808-utilities\") pod \"ae7aaa0b-3da0-4764-808a-619dae2fe808\" (UID: \"ae7aaa0b-3da0-4764-808a-619dae2fe808\") " Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.687901 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8951296-b323-4ac3-9e01-963b60ff998c-catalog-content\") pod \"f8951296-b323-4ac3-9e01-963b60ff998c\" (UID: \"f8951296-b323-4ac3-9e01-963b60ff998c\") " Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.687920 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae7aaa0b-3da0-4764-808a-619dae2fe808-catalog-content\") pod \"ae7aaa0b-3da0-4764-808a-619dae2fe808\" (UID: \"ae7aaa0b-3da0-4764-808a-619dae2fe808\") " Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.687946 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hk262\" (UniqueName: \"kubernetes.io/projected/1383ed81-5c2a-4416-8c67-05e4881db7be-kube-api-access-hk262\") pod \"1383ed81-5c2a-4416-8c67-05e4881db7be\" (UID: \"1383ed81-5c2a-4416-8c67-05e4881db7be\") " Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.688004 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b08de060-f5b7-4bed-a27c-89294a8827fa-marketplace-trusted-ca\") pod \"b08de060-f5b7-4bed-a27c-89294a8827fa\" (UID: \"b08de060-f5b7-4bed-a27c-89294a8827fa\") " Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.688028 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1383ed81-5c2a-4416-8c67-05e4881db7be-catalog-content\") pod \"1383ed81-5c2a-4416-8c67-05e4881db7be\" (UID: \"1383ed81-5c2a-4416-8c67-05e4881db7be\") " Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.688044 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c707d1d-d2c2-42db-82b9-afe8fde54914-catalog-content\") pod \"9c707d1d-d2c2-42db-82b9-afe8fde54914\" (UID: \"9c707d1d-d2c2-42db-82b9-afe8fde54914\") " Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.688076 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1383ed81-5c2a-4416-8c67-05e4881db7be-utilities\") pod \"1383ed81-5c2a-4416-8c67-05e4881db7be\" (UID: \"1383ed81-5c2a-4416-8c67-05e4881db7be\") " Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.688099 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drtpr\" (UniqueName: \"kubernetes.io/projected/9c707d1d-d2c2-42db-82b9-afe8fde54914-kube-api-access-drtpr\") pod \"9c707d1d-d2c2-42db-82b9-afe8fde54914\" (UID: \"9c707d1d-d2c2-42db-82b9-afe8fde54914\") " Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.688116 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58z8t\" (UniqueName: \"kubernetes.io/projected/f8951296-b323-4ac3-9e01-963b60ff998c-kube-api-access-58z8t\") pod \"f8951296-b323-4ac3-9e01-963b60ff998c\" (UID: \"f8951296-b323-4ac3-9e01-963b60ff998c\") " Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.688135 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b08de060-f5b7-4bed-a27c-89294a8827fa-marketplace-operator-metrics\") pod \"b08de060-f5b7-4bed-a27c-89294a8827fa\" (UID: \"b08de060-f5b7-4bed-a27c-89294a8827fa\") " Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.688149 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8951296-b323-4ac3-9e01-963b60ff998c-utilities\") pod \"f8951296-b323-4ac3-9e01-963b60ff998c\" (UID: \"f8951296-b323-4ac3-9e01-963b60ff998c\") " Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.688169 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5qpl\" (UniqueName: \"kubernetes.io/projected/b08de060-f5b7-4bed-a27c-89294a8827fa-kube-api-access-j5qpl\") pod \"b08de060-f5b7-4bed-a27c-89294a8827fa\" (UID: \"b08de060-f5b7-4bed-a27c-89294a8827fa\") " Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.688183 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6kpl\" (UniqueName: \"kubernetes.io/projected/ae7aaa0b-3da0-4764-808a-619dae2fe808-kube-api-access-j6kpl\") pod \"ae7aaa0b-3da0-4764-808a-619dae2fe808\" (UID: \"ae7aaa0b-3da0-4764-808a-619dae2fe808\") " Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.688207 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c707d1d-d2c2-42db-82b9-afe8fde54914-utilities\") pod \"9c707d1d-d2c2-42db-82b9-afe8fde54914\" (UID: \"9c707d1d-d2c2-42db-82b9-afe8fde54914\") " Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.690215 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1383ed81-5c2a-4416-8c67-05e4881db7be-utilities" (OuterVolumeSpecName: "utilities") pod "1383ed81-5c2a-4416-8c67-05e4881db7be" (UID: "1383ed81-5c2a-4416-8c67-05e4881db7be"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.694142 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b08de060-f5b7-4bed-a27c-89294a8827fa-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b08de060-f5b7-4bed-a27c-89294a8827fa" (UID: "b08de060-f5b7-4bed-a27c-89294a8827fa"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.696901 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c707d1d-d2c2-42db-82b9-afe8fde54914-utilities" (OuterVolumeSpecName: "utilities") pod "9c707d1d-d2c2-42db-82b9-afe8fde54914" (UID: "9c707d1d-d2c2-42db-82b9-afe8fde54914"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.697462 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c707d1d-d2c2-42db-82b9-afe8fde54914-kube-api-access-drtpr" (OuterVolumeSpecName: "kube-api-access-drtpr") pod "9c707d1d-d2c2-42db-82b9-afe8fde54914" (UID: "9c707d1d-d2c2-42db-82b9-afe8fde54914"). InnerVolumeSpecName "kube-api-access-drtpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.698219 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b08de060-f5b7-4bed-a27c-89294a8827fa-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b08de060-f5b7-4bed-a27c-89294a8827fa" (UID: "b08de060-f5b7-4bed-a27c-89294a8827fa"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.698324 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae7aaa0b-3da0-4764-808a-619dae2fe808-utilities" (OuterVolumeSpecName: "utilities") pod "ae7aaa0b-3da0-4764-808a-619dae2fe808" (UID: "ae7aaa0b-3da0-4764-808a-619dae2fe808"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.699191 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8951296-b323-4ac3-9e01-963b60ff998c-utilities" (OuterVolumeSpecName: "utilities") pod "f8951296-b323-4ac3-9e01-963b60ff998c" (UID: "f8951296-b323-4ac3-9e01-963b60ff998c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.700402 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b08de060-f5b7-4bed-a27c-89294a8827fa-kube-api-access-j5qpl" (OuterVolumeSpecName: "kube-api-access-j5qpl") pod "b08de060-f5b7-4bed-a27c-89294a8827fa" (UID: "b08de060-f5b7-4bed-a27c-89294a8827fa"). InnerVolumeSpecName "kube-api-access-j5qpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.701210 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae7aaa0b-3da0-4764-808a-619dae2fe808-kube-api-access-j6kpl" (OuterVolumeSpecName: "kube-api-access-j6kpl") pod "ae7aaa0b-3da0-4764-808a-619dae2fe808" (UID: "ae7aaa0b-3da0-4764-808a-619dae2fe808"). InnerVolumeSpecName "kube-api-access-j6kpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.702282 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8951296-b323-4ac3-9e01-963b60ff998c-kube-api-access-58z8t" (OuterVolumeSpecName: "kube-api-access-58z8t") pod "f8951296-b323-4ac3-9e01-963b60ff998c" (UID: "f8951296-b323-4ac3-9e01-963b60ff998c"). InnerVolumeSpecName "kube-api-access-58z8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.728868 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1383ed81-5c2a-4416-8c67-05e4881db7be-kube-api-access-hk262" (OuterVolumeSpecName: "kube-api-access-hk262") pod "1383ed81-5c2a-4416-8c67-05e4881db7be" (UID: "1383ed81-5c2a-4416-8c67-05e4881db7be"). InnerVolumeSpecName "kube-api-access-hk262". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.735720 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c707d1d-d2c2-42db-82b9-afe8fde54914-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9c707d1d-d2c2-42db-82b9-afe8fde54914" (UID: "9c707d1d-d2c2-42db-82b9-afe8fde54914"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.765576 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1383ed81-5c2a-4416-8c67-05e4881db7be-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1383ed81-5c2a-4416-8c67-05e4881db7be" (UID: "1383ed81-5c2a-4416-8c67-05e4881db7be"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.789219 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae7aaa0b-3da0-4764-808a-619dae2fe808-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.789258 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hk262\" (UniqueName: \"kubernetes.io/projected/1383ed81-5c2a-4416-8c67-05e4881db7be-kube-api-access-hk262\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.789271 4809 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b08de060-f5b7-4bed-a27c-89294a8827fa-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.789283 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1383ed81-5c2a-4416-8c67-05e4881db7be-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.789295 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c707d1d-d2c2-42db-82b9-afe8fde54914-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.789305 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1383ed81-5c2a-4416-8c67-05e4881db7be-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.789316 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drtpr\" (UniqueName: \"kubernetes.io/projected/9c707d1d-d2c2-42db-82b9-afe8fde54914-kube-api-access-drtpr\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.789327 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58z8t\" (UniqueName: \"kubernetes.io/projected/f8951296-b323-4ac3-9e01-963b60ff998c-kube-api-access-58z8t\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.789338 4809 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b08de060-f5b7-4bed-a27c-89294a8827fa-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.789349 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8951296-b323-4ac3-9e01-963b60ff998c-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.789360 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5qpl\" (UniqueName: \"kubernetes.io/projected/b08de060-f5b7-4bed-a27c-89294a8827fa-kube-api-access-j5qpl\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.789371 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6kpl\" (UniqueName: \"kubernetes.io/projected/ae7aaa0b-3da0-4764-808a-619dae2fe808-kube-api-access-j6kpl\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.789382 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c707d1d-d2c2-42db-82b9-afe8fde54914-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.808608 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8951296-b323-4ac3-9e01-963b60ff998c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8951296-b323-4ac3-9e01-963b60ff998c" (UID: "f8951296-b323-4ac3-9e01-963b60ff998c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.866448 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae7aaa0b-3da0-4764-808a-619dae2fe808-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae7aaa0b-3da0-4764-808a-619dae2fe808" (UID: "ae7aaa0b-3da0-4764-808a-619dae2fe808"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.890659 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8951296-b323-4ac3-9e01-963b60ff998c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.890694 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae7aaa0b-3da0-4764-808a-619dae2fe808-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.959989 4809 generic.go:334] "Generic (PLEG): container finished" podID="f8951296-b323-4ac3-9e01-963b60ff998c" containerID="05d73028a172bc3b0d3cd3d56d5f7ca20031d690d99abc7f0ed30e121e7592f4" exitCode=0 Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.960060 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llxlp" event={"ID":"f8951296-b323-4ac3-9e01-963b60ff998c","Type":"ContainerDied","Data":"05d73028a172bc3b0d3cd3d56d5f7ca20031d690d99abc7f0ed30e121e7592f4"} Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.960133 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llxlp" event={"ID":"f8951296-b323-4ac3-9e01-963b60ff998c","Type":"ContainerDied","Data":"10efd5615b838c8f298a6c3da72c970496b9cb66d0fcbfc009a579b7e4164ff0"} Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.960073 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-llxlp" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.960157 4809 scope.go:117] "RemoveContainer" containerID="05d73028a172bc3b0d3cd3d56d5f7ca20031d690d99abc7f0ed30e121e7592f4" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.962653 4809 generic.go:334] "Generic (PLEG): container finished" podID="9c707d1d-d2c2-42db-82b9-afe8fde54914" containerID="1b606756ec79d1f682e5df4032fefde8a8a0aac4d23d4446bf79e9e28d8d07a1" exitCode=0 Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.962731 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9r2kr" event={"ID":"9c707d1d-d2c2-42db-82b9-afe8fde54914","Type":"ContainerDied","Data":"1b606756ec79d1f682e5df4032fefde8a8a0aac4d23d4446bf79e9e28d8d07a1"} Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.962759 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9r2kr" event={"ID":"9c707d1d-d2c2-42db-82b9-afe8fde54914","Type":"ContainerDied","Data":"8f7718e2a085edd75d56c3555f57acab01e0fb29524f3c74b10aad7f06cdde4b"} Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.962829 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9r2kr" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.965564 4809 generic.go:334] "Generic (PLEG): container finished" podID="1383ed81-5c2a-4416-8c67-05e4881db7be" containerID="f69124e68a7c6e2ee6c394c6f522c6b844a3e301c21c65411b5d3be131f02e34" exitCode=0 Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.965631 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pj46g" event={"ID":"1383ed81-5c2a-4416-8c67-05e4881db7be","Type":"ContainerDied","Data":"f69124e68a7c6e2ee6c394c6f522c6b844a3e301c21c65411b5d3be131f02e34"} Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.965665 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pj46g" event={"ID":"1383ed81-5c2a-4416-8c67-05e4881db7be","Type":"ContainerDied","Data":"66c7cd27d613d48d506293c04b3f19db0acfa6edb3398e30423b3945cb41f9a3"} Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.965980 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pj46g" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.969077 4809 generic.go:334] "Generic (PLEG): container finished" podID="ae7aaa0b-3da0-4764-808a-619dae2fe808" containerID="11d2beca7140e69ba1bd7817a34ca8e127d922b36d5605c08aa05deb627baea0" exitCode=0 Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.969117 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wsc5l" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.969138 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsc5l" event={"ID":"ae7aaa0b-3da0-4764-808a-619dae2fe808","Type":"ContainerDied","Data":"11d2beca7140e69ba1bd7817a34ca8e127d922b36d5605c08aa05deb627baea0"} Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.969169 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsc5l" event={"ID":"ae7aaa0b-3da0-4764-808a-619dae2fe808","Type":"ContainerDied","Data":"b4355799703c087442c717fe6dec81edcb8f451852047a00f6d985e696796533"} Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.971674 4809 generic.go:334] "Generic (PLEG): container finished" podID="b08de060-f5b7-4bed-a27c-89294a8827fa" containerID="f5b1bb43ee1cfb566a2b2292ef38df7c63aeaee858c08fee970d588092fb019e" exitCode=0 Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.971704 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.971728 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" event={"ID":"b08de060-f5b7-4bed-a27c-89294a8827fa","Type":"ContainerDied","Data":"f5b1bb43ee1cfb566a2b2292ef38df7c63aeaee858c08fee970d588092fb019e"} Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.971870 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-svxc5" event={"ID":"b08de060-f5b7-4bed-a27c-89294a8827fa","Type":"ContainerDied","Data":"053b70f21e3f4b82b650ef00d8249129b288d2a2f87131adf31e03b86cc0ca72"} Nov 24 06:59:27 crc kubenswrapper[4809]: I1124 06:59:27.975456 4809 scope.go:117] "RemoveContainer" containerID="abeacc8c9f08cd94466ce62e0c61f64d136cc0e42176bff9132abc9f7ef03c83" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.015013 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-llxlp"] Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.016618 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-llxlp"] Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.020239 4809 scope.go:117] "RemoveContainer" containerID="064add890a06df5449f17b0d1227089f5f034abca0fc1bdae562272bd597ba2d" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.033917 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9r2kr"] Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.036927 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9r2kr"] Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.037114 4809 scope.go:117] "RemoveContainer" containerID="05d73028a172bc3b0d3cd3d56d5f7ca20031d690d99abc7f0ed30e121e7592f4" Nov 24 06:59:28 crc kubenswrapper[4809]: E1124 06:59:28.037461 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05d73028a172bc3b0d3cd3d56d5f7ca20031d690d99abc7f0ed30e121e7592f4\": container with ID starting with 05d73028a172bc3b0d3cd3d56d5f7ca20031d690d99abc7f0ed30e121e7592f4 not found: ID does not exist" containerID="05d73028a172bc3b0d3cd3d56d5f7ca20031d690d99abc7f0ed30e121e7592f4" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.037509 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05d73028a172bc3b0d3cd3d56d5f7ca20031d690d99abc7f0ed30e121e7592f4"} err="failed to get container status \"05d73028a172bc3b0d3cd3d56d5f7ca20031d690d99abc7f0ed30e121e7592f4\": rpc error: code = NotFound desc = could not find container \"05d73028a172bc3b0d3cd3d56d5f7ca20031d690d99abc7f0ed30e121e7592f4\": container with ID starting with 05d73028a172bc3b0d3cd3d56d5f7ca20031d690d99abc7f0ed30e121e7592f4 not found: ID does not exist" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.037534 4809 scope.go:117] "RemoveContainer" containerID="abeacc8c9f08cd94466ce62e0c61f64d136cc0e42176bff9132abc9f7ef03c83" Nov 24 06:59:28 crc kubenswrapper[4809]: E1124 06:59:28.038051 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abeacc8c9f08cd94466ce62e0c61f64d136cc0e42176bff9132abc9f7ef03c83\": container with ID starting with abeacc8c9f08cd94466ce62e0c61f64d136cc0e42176bff9132abc9f7ef03c83 not found: ID does not exist" containerID="abeacc8c9f08cd94466ce62e0c61f64d136cc0e42176bff9132abc9f7ef03c83" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.038090 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abeacc8c9f08cd94466ce62e0c61f64d136cc0e42176bff9132abc9f7ef03c83"} err="failed to get container status \"abeacc8c9f08cd94466ce62e0c61f64d136cc0e42176bff9132abc9f7ef03c83\": rpc error: code = NotFound desc = could not find container \"abeacc8c9f08cd94466ce62e0c61f64d136cc0e42176bff9132abc9f7ef03c83\": container with ID starting with abeacc8c9f08cd94466ce62e0c61f64d136cc0e42176bff9132abc9f7ef03c83 not found: ID does not exist" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.038104 4809 scope.go:117] "RemoveContainer" containerID="064add890a06df5449f17b0d1227089f5f034abca0fc1bdae562272bd597ba2d" Nov 24 06:59:28 crc kubenswrapper[4809]: E1124 06:59:28.040191 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"064add890a06df5449f17b0d1227089f5f034abca0fc1bdae562272bd597ba2d\": container with ID starting with 064add890a06df5449f17b0d1227089f5f034abca0fc1bdae562272bd597ba2d not found: ID does not exist" containerID="064add890a06df5449f17b0d1227089f5f034abca0fc1bdae562272bd597ba2d" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.040230 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"064add890a06df5449f17b0d1227089f5f034abca0fc1bdae562272bd597ba2d"} err="failed to get container status \"064add890a06df5449f17b0d1227089f5f034abca0fc1bdae562272bd597ba2d\": rpc error: code = NotFound desc = could not find container \"064add890a06df5449f17b0d1227089f5f034abca0fc1bdae562272bd597ba2d\": container with ID starting with 064add890a06df5449f17b0d1227089f5f034abca0fc1bdae562272bd597ba2d not found: ID does not exist" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.040256 4809 scope.go:117] "RemoveContainer" containerID="1b606756ec79d1f682e5df4032fefde8a8a0aac4d23d4446bf79e9e28d8d07a1" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.044484 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-svxc5"] Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.048329 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-svxc5"] Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.061781 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zqwx5"] Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.062456 4809 scope.go:117] "RemoveContainer" containerID="cae761370ec4076c7c309bb9c17191b404ce1365f20d1659b4595908a484cd10" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.068072 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wsc5l"] Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.070709 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wsc5l"] Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.074984 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pj46g"] Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.081152 4809 scope.go:117] "RemoveContainer" containerID="56f4369c77a19d497136aba158c1b1e368e3fc74937c1cbbc1f4a5e5ea52a22d" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.083665 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pj46g"] Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.098210 4809 scope.go:117] "RemoveContainer" containerID="1b606756ec79d1f682e5df4032fefde8a8a0aac4d23d4446bf79e9e28d8d07a1" Nov 24 06:59:28 crc kubenswrapper[4809]: E1124 06:59:28.098628 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b606756ec79d1f682e5df4032fefde8a8a0aac4d23d4446bf79e9e28d8d07a1\": container with ID starting with 1b606756ec79d1f682e5df4032fefde8a8a0aac4d23d4446bf79e9e28d8d07a1 not found: ID does not exist" containerID="1b606756ec79d1f682e5df4032fefde8a8a0aac4d23d4446bf79e9e28d8d07a1" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.098655 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b606756ec79d1f682e5df4032fefde8a8a0aac4d23d4446bf79e9e28d8d07a1"} err="failed to get container status \"1b606756ec79d1f682e5df4032fefde8a8a0aac4d23d4446bf79e9e28d8d07a1\": rpc error: code = NotFound desc = could not find container \"1b606756ec79d1f682e5df4032fefde8a8a0aac4d23d4446bf79e9e28d8d07a1\": container with ID starting with 1b606756ec79d1f682e5df4032fefde8a8a0aac4d23d4446bf79e9e28d8d07a1 not found: ID does not exist" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.098677 4809 scope.go:117] "RemoveContainer" containerID="cae761370ec4076c7c309bb9c17191b404ce1365f20d1659b4595908a484cd10" Nov 24 06:59:28 crc kubenswrapper[4809]: E1124 06:59:28.099093 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cae761370ec4076c7c309bb9c17191b404ce1365f20d1659b4595908a484cd10\": container with ID starting with cae761370ec4076c7c309bb9c17191b404ce1365f20d1659b4595908a484cd10 not found: ID does not exist" containerID="cae761370ec4076c7c309bb9c17191b404ce1365f20d1659b4595908a484cd10" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.099113 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cae761370ec4076c7c309bb9c17191b404ce1365f20d1659b4595908a484cd10"} err="failed to get container status \"cae761370ec4076c7c309bb9c17191b404ce1365f20d1659b4595908a484cd10\": rpc error: code = NotFound desc = could not find container \"cae761370ec4076c7c309bb9c17191b404ce1365f20d1659b4595908a484cd10\": container with ID starting with cae761370ec4076c7c309bb9c17191b404ce1365f20d1659b4595908a484cd10 not found: ID does not exist" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.099129 4809 scope.go:117] "RemoveContainer" containerID="56f4369c77a19d497136aba158c1b1e368e3fc74937c1cbbc1f4a5e5ea52a22d" Nov 24 06:59:28 crc kubenswrapper[4809]: E1124 06:59:28.099330 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56f4369c77a19d497136aba158c1b1e368e3fc74937c1cbbc1f4a5e5ea52a22d\": container with ID starting with 56f4369c77a19d497136aba158c1b1e368e3fc74937c1cbbc1f4a5e5ea52a22d not found: ID does not exist" containerID="56f4369c77a19d497136aba158c1b1e368e3fc74937c1cbbc1f4a5e5ea52a22d" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.099351 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56f4369c77a19d497136aba158c1b1e368e3fc74937c1cbbc1f4a5e5ea52a22d"} err="failed to get container status \"56f4369c77a19d497136aba158c1b1e368e3fc74937c1cbbc1f4a5e5ea52a22d\": rpc error: code = NotFound desc = could not find container \"56f4369c77a19d497136aba158c1b1e368e3fc74937c1cbbc1f4a5e5ea52a22d\": container with ID starting with 56f4369c77a19d497136aba158c1b1e368e3fc74937c1cbbc1f4a5e5ea52a22d not found: ID does not exist" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.099368 4809 scope.go:117] "RemoveContainer" containerID="f69124e68a7c6e2ee6c394c6f522c6b844a3e301c21c65411b5d3be131f02e34" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.111234 4809 scope.go:117] "RemoveContainer" containerID="b46af9efba74659737669b84699e3efa267ab313b341e0e4fd52a019312e2221" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.134189 4809 scope.go:117] "RemoveContainer" containerID="d41fd86b045fd2a17480ad649f2ddc752efbf24b64eb1325608f4f1c9bc9b5ae" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.152448 4809 scope.go:117] "RemoveContainer" containerID="f69124e68a7c6e2ee6c394c6f522c6b844a3e301c21c65411b5d3be131f02e34" Nov 24 06:59:28 crc kubenswrapper[4809]: E1124 06:59:28.153767 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f69124e68a7c6e2ee6c394c6f522c6b844a3e301c21c65411b5d3be131f02e34\": container with ID starting with f69124e68a7c6e2ee6c394c6f522c6b844a3e301c21c65411b5d3be131f02e34 not found: ID does not exist" containerID="f69124e68a7c6e2ee6c394c6f522c6b844a3e301c21c65411b5d3be131f02e34" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.153828 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f69124e68a7c6e2ee6c394c6f522c6b844a3e301c21c65411b5d3be131f02e34"} err="failed to get container status \"f69124e68a7c6e2ee6c394c6f522c6b844a3e301c21c65411b5d3be131f02e34\": rpc error: code = NotFound desc = could not find container \"f69124e68a7c6e2ee6c394c6f522c6b844a3e301c21c65411b5d3be131f02e34\": container with ID starting with f69124e68a7c6e2ee6c394c6f522c6b844a3e301c21c65411b5d3be131f02e34 not found: ID does not exist" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.153892 4809 scope.go:117] "RemoveContainer" containerID="b46af9efba74659737669b84699e3efa267ab313b341e0e4fd52a019312e2221" Nov 24 06:59:28 crc kubenswrapper[4809]: E1124 06:59:28.154746 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b46af9efba74659737669b84699e3efa267ab313b341e0e4fd52a019312e2221\": container with ID starting with b46af9efba74659737669b84699e3efa267ab313b341e0e4fd52a019312e2221 not found: ID does not exist" containerID="b46af9efba74659737669b84699e3efa267ab313b341e0e4fd52a019312e2221" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.154776 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b46af9efba74659737669b84699e3efa267ab313b341e0e4fd52a019312e2221"} err="failed to get container status \"b46af9efba74659737669b84699e3efa267ab313b341e0e4fd52a019312e2221\": rpc error: code = NotFound desc = could not find container \"b46af9efba74659737669b84699e3efa267ab313b341e0e4fd52a019312e2221\": container with ID starting with b46af9efba74659737669b84699e3efa267ab313b341e0e4fd52a019312e2221 not found: ID does not exist" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.154797 4809 scope.go:117] "RemoveContainer" containerID="d41fd86b045fd2a17480ad649f2ddc752efbf24b64eb1325608f4f1c9bc9b5ae" Nov 24 06:59:28 crc kubenswrapper[4809]: E1124 06:59:28.155258 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d41fd86b045fd2a17480ad649f2ddc752efbf24b64eb1325608f4f1c9bc9b5ae\": container with ID starting with d41fd86b045fd2a17480ad649f2ddc752efbf24b64eb1325608f4f1c9bc9b5ae not found: ID does not exist" containerID="d41fd86b045fd2a17480ad649f2ddc752efbf24b64eb1325608f4f1c9bc9b5ae" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.155287 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d41fd86b045fd2a17480ad649f2ddc752efbf24b64eb1325608f4f1c9bc9b5ae"} err="failed to get container status \"d41fd86b045fd2a17480ad649f2ddc752efbf24b64eb1325608f4f1c9bc9b5ae\": rpc error: code = NotFound desc = could not find container \"d41fd86b045fd2a17480ad649f2ddc752efbf24b64eb1325608f4f1c9bc9b5ae\": container with ID starting with d41fd86b045fd2a17480ad649f2ddc752efbf24b64eb1325608f4f1c9bc9b5ae not found: ID does not exist" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.155306 4809 scope.go:117] "RemoveContainer" containerID="11d2beca7140e69ba1bd7817a34ca8e127d922b36d5605c08aa05deb627baea0" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.173447 4809 scope.go:117] "RemoveContainer" containerID="22bcfb0f275251f85af8aadda4c68d69015b9f71e19a4c45b74c420b8838d835" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.190501 4809 scope.go:117] "RemoveContainer" containerID="9275f80fce7ee8d0fcacf680bf7f71ee2ae6f6be736e2da04beaeb19bc7d332b" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.209892 4809 scope.go:117] "RemoveContainer" containerID="11d2beca7140e69ba1bd7817a34ca8e127d922b36d5605c08aa05deb627baea0" Nov 24 06:59:28 crc kubenswrapper[4809]: E1124 06:59:28.210598 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11d2beca7140e69ba1bd7817a34ca8e127d922b36d5605c08aa05deb627baea0\": container with ID starting with 11d2beca7140e69ba1bd7817a34ca8e127d922b36d5605c08aa05deb627baea0 not found: ID does not exist" containerID="11d2beca7140e69ba1bd7817a34ca8e127d922b36d5605c08aa05deb627baea0" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.210637 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11d2beca7140e69ba1bd7817a34ca8e127d922b36d5605c08aa05deb627baea0"} err="failed to get container status \"11d2beca7140e69ba1bd7817a34ca8e127d922b36d5605c08aa05deb627baea0\": rpc error: code = NotFound desc = could not find container \"11d2beca7140e69ba1bd7817a34ca8e127d922b36d5605c08aa05deb627baea0\": container with ID starting with 11d2beca7140e69ba1bd7817a34ca8e127d922b36d5605c08aa05deb627baea0 not found: ID does not exist" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.210665 4809 scope.go:117] "RemoveContainer" containerID="22bcfb0f275251f85af8aadda4c68d69015b9f71e19a4c45b74c420b8838d835" Nov 24 06:59:28 crc kubenswrapper[4809]: E1124 06:59:28.211439 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22bcfb0f275251f85af8aadda4c68d69015b9f71e19a4c45b74c420b8838d835\": container with ID starting with 22bcfb0f275251f85af8aadda4c68d69015b9f71e19a4c45b74c420b8838d835 not found: ID does not exist" containerID="22bcfb0f275251f85af8aadda4c68d69015b9f71e19a4c45b74c420b8838d835" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.211469 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22bcfb0f275251f85af8aadda4c68d69015b9f71e19a4c45b74c420b8838d835"} err="failed to get container status \"22bcfb0f275251f85af8aadda4c68d69015b9f71e19a4c45b74c420b8838d835\": rpc error: code = NotFound desc = could not find container \"22bcfb0f275251f85af8aadda4c68d69015b9f71e19a4c45b74c420b8838d835\": container with ID starting with 22bcfb0f275251f85af8aadda4c68d69015b9f71e19a4c45b74c420b8838d835 not found: ID does not exist" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.211493 4809 scope.go:117] "RemoveContainer" containerID="9275f80fce7ee8d0fcacf680bf7f71ee2ae6f6be736e2da04beaeb19bc7d332b" Nov 24 06:59:28 crc kubenswrapper[4809]: E1124 06:59:28.211902 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9275f80fce7ee8d0fcacf680bf7f71ee2ae6f6be736e2da04beaeb19bc7d332b\": container with ID starting with 9275f80fce7ee8d0fcacf680bf7f71ee2ae6f6be736e2da04beaeb19bc7d332b not found: ID does not exist" containerID="9275f80fce7ee8d0fcacf680bf7f71ee2ae6f6be736e2da04beaeb19bc7d332b" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.211932 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9275f80fce7ee8d0fcacf680bf7f71ee2ae6f6be736e2da04beaeb19bc7d332b"} err="failed to get container status \"9275f80fce7ee8d0fcacf680bf7f71ee2ae6f6be736e2da04beaeb19bc7d332b\": rpc error: code = NotFound desc = could not find container \"9275f80fce7ee8d0fcacf680bf7f71ee2ae6f6be736e2da04beaeb19bc7d332b\": container with ID starting with 9275f80fce7ee8d0fcacf680bf7f71ee2ae6f6be736e2da04beaeb19bc7d332b not found: ID does not exist" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.211948 4809 scope.go:117] "RemoveContainer" containerID="f5b1bb43ee1cfb566a2b2292ef38df7c63aeaee858c08fee970d588092fb019e" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.228343 4809 scope.go:117] "RemoveContainer" containerID="f5b1bb43ee1cfb566a2b2292ef38df7c63aeaee858c08fee970d588092fb019e" Nov 24 06:59:28 crc kubenswrapper[4809]: E1124 06:59:28.228838 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5b1bb43ee1cfb566a2b2292ef38df7c63aeaee858c08fee970d588092fb019e\": container with ID starting with f5b1bb43ee1cfb566a2b2292ef38df7c63aeaee858c08fee970d588092fb019e not found: ID does not exist" containerID="f5b1bb43ee1cfb566a2b2292ef38df7c63aeaee858c08fee970d588092fb019e" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.228868 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5b1bb43ee1cfb566a2b2292ef38df7c63aeaee858c08fee970d588092fb019e"} err="failed to get container status \"f5b1bb43ee1cfb566a2b2292ef38df7c63aeaee858c08fee970d588092fb019e\": rpc error: code = NotFound desc = could not find container \"f5b1bb43ee1cfb566a2b2292ef38df7c63aeaee858c08fee970d588092fb019e\": container with ID starting with f5b1bb43ee1cfb566a2b2292ef38df7c63aeaee858c08fee970d588092fb019e not found: ID does not exist" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.898694 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1383ed81-5c2a-4416-8c67-05e4881db7be" path="/var/lib/kubelet/pods/1383ed81-5c2a-4416-8c67-05e4881db7be/volumes" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.899990 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c707d1d-d2c2-42db-82b9-afe8fde54914" path="/var/lib/kubelet/pods/9c707d1d-d2c2-42db-82b9-afe8fde54914/volumes" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.900793 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae7aaa0b-3da0-4764-808a-619dae2fe808" path="/var/lib/kubelet/pods/ae7aaa0b-3da0-4764-808a-619dae2fe808/volumes" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.902597 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b08de060-f5b7-4bed-a27c-89294a8827fa" path="/var/lib/kubelet/pods/b08de060-f5b7-4bed-a27c-89294a8827fa/volumes" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.903292 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8951296-b323-4ac3-9e01-963b60ff998c" path="/var/lib/kubelet/pods/f8951296-b323-4ac3-9e01-963b60ff998c/volumes" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.982600 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zqwx5" event={"ID":"4b7632a5-d284-4a30-89dc-c301a7d13609","Type":"ContainerStarted","Data":"be23879e7d53c2fb28f6aa00ee35c763eced74d88fff50c7e9bca7022a636c84"} Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.982641 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zqwx5" event={"ID":"4b7632a5-d284-4a30-89dc-c301a7d13609","Type":"ContainerStarted","Data":"168b04246b1ed0a25ded2888c45a527d89446a3ca5998528b6362da1d01dc576"} Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.984914 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-zqwx5" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.987485 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-zqwx5" Nov 24 06:59:28 crc kubenswrapper[4809]: I1124 06:59:28.998346 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-zqwx5" podStartSLOduration=1.998316654 podStartE2EDuration="1.998316654s" podCreationTimestamp="2025-11-24 06:59:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:59:28.998084927 +0000 UTC m=+268.898676632" watchObservedRunningTime="2025-11-24 06:59:28.998316654 +0000 UTC m=+268.898908359" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.366439 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-l9v7r"] Nov 24 06:59:29 crc kubenswrapper[4809]: E1124 06:59:29.366613 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae7aaa0b-3da0-4764-808a-619dae2fe808" containerName="extract-utilities" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.366624 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae7aaa0b-3da0-4764-808a-619dae2fe808" containerName="extract-utilities" Nov 24 06:59:29 crc kubenswrapper[4809]: E1124 06:59:29.366632 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae7aaa0b-3da0-4764-808a-619dae2fe808" containerName="extract-content" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.366638 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae7aaa0b-3da0-4764-808a-619dae2fe808" containerName="extract-content" Nov 24 06:59:29 crc kubenswrapper[4809]: E1124 06:59:29.366648 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1383ed81-5c2a-4416-8c67-05e4881db7be" containerName="extract-content" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.366654 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1383ed81-5c2a-4416-8c67-05e4881db7be" containerName="extract-content" Nov 24 06:59:29 crc kubenswrapper[4809]: E1124 06:59:29.366665 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b08de060-f5b7-4bed-a27c-89294a8827fa" containerName="marketplace-operator" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.366671 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b08de060-f5b7-4bed-a27c-89294a8827fa" containerName="marketplace-operator" Nov 24 06:59:29 crc kubenswrapper[4809]: E1124 06:59:29.366679 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8951296-b323-4ac3-9e01-963b60ff998c" containerName="extract-utilities" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.366685 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8951296-b323-4ac3-9e01-963b60ff998c" containerName="extract-utilities" Nov 24 06:59:29 crc kubenswrapper[4809]: E1124 06:59:29.366694 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1383ed81-5c2a-4416-8c67-05e4881db7be" containerName="extract-utilities" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.366700 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1383ed81-5c2a-4416-8c67-05e4881db7be" containerName="extract-utilities" Nov 24 06:59:29 crc kubenswrapper[4809]: E1124 06:59:29.366708 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8951296-b323-4ac3-9e01-963b60ff998c" containerName="registry-server" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.366713 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8951296-b323-4ac3-9e01-963b60ff998c" containerName="registry-server" Nov 24 06:59:29 crc kubenswrapper[4809]: E1124 06:59:29.366723 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c707d1d-d2c2-42db-82b9-afe8fde54914" containerName="extract-utilities" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.366728 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c707d1d-d2c2-42db-82b9-afe8fde54914" containerName="extract-utilities" Nov 24 06:59:29 crc kubenswrapper[4809]: E1124 06:59:29.366734 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c707d1d-d2c2-42db-82b9-afe8fde54914" containerName="extract-content" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.366740 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c707d1d-d2c2-42db-82b9-afe8fde54914" containerName="extract-content" Nov 24 06:59:29 crc kubenswrapper[4809]: E1124 06:59:29.366747 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1383ed81-5c2a-4416-8c67-05e4881db7be" containerName="registry-server" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.366753 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1383ed81-5c2a-4416-8c67-05e4881db7be" containerName="registry-server" Nov 24 06:59:29 crc kubenswrapper[4809]: E1124 06:59:29.366761 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae7aaa0b-3da0-4764-808a-619dae2fe808" containerName="registry-server" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.366766 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae7aaa0b-3da0-4764-808a-619dae2fe808" containerName="registry-server" Nov 24 06:59:29 crc kubenswrapper[4809]: E1124 06:59:29.366777 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c707d1d-d2c2-42db-82b9-afe8fde54914" containerName="registry-server" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.366782 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c707d1d-d2c2-42db-82b9-afe8fde54914" containerName="registry-server" Nov 24 06:59:29 crc kubenswrapper[4809]: E1124 06:59:29.366789 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8951296-b323-4ac3-9e01-963b60ff998c" containerName="extract-content" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.366794 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8951296-b323-4ac3-9e01-963b60ff998c" containerName="extract-content" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.366878 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae7aaa0b-3da0-4764-808a-619dae2fe808" containerName="registry-server" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.366888 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8951296-b323-4ac3-9e01-963b60ff998c" containerName="registry-server" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.366899 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c707d1d-d2c2-42db-82b9-afe8fde54914" containerName="registry-server" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.366910 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1383ed81-5c2a-4416-8c67-05e4881db7be" containerName="registry-server" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.366917 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b08de060-f5b7-4bed-a27c-89294a8827fa" containerName="marketplace-operator" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.367540 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l9v7r" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.369061 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.383668 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l9v7r"] Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.410914 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9586cac6-9d7e-4c04-9400-b0760540e65c-utilities\") pod \"redhat-marketplace-l9v7r\" (UID: \"9586cac6-9d7e-4c04-9400-b0760540e65c\") " pod="openshift-marketplace/redhat-marketplace-l9v7r" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.411018 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bt4zh\" (UniqueName: \"kubernetes.io/projected/9586cac6-9d7e-4c04-9400-b0760540e65c-kube-api-access-bt4zh\") pod \"redhat-marketplace-l9v7r\" (UID: \"9586cac6-9d7e-4c04-9400-b0760540e65c\") " pod="openshift-marketplace/redhat-marketplace-l9v7r" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.411038 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9586cac6-9d7e-4c04-9400-b0760540e65c-catalog-content\") pod \"redhat-marketplace-l9v7r\" (UID: \"9586cac6-9d7e-4c04-9400-b0760540e65c\") " pod="openshift-marketplace/redhat-marketplace-l9v7r" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.511428 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9586cac6-9d7e-4c04-9400-b0760540e65c-catalog-content\") pod \"redhat-marketplace-l9v7r\" (UID: \"9586cac6-9d7e-4c04-9400-b0760540e65c\") " pod="openshift-marketplace/redhat-marketplace-l9v7r" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.511490 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9586cac6-9d7e-4c04-9400-b0760540e65c-utilities\") pod \"redhat-marketplace-l9v7r\" (UID: \"9586cac6-9d7e-4c04-9400-b0760540e65c\") " pod="openshift-marketplace/redhat-marketplace-l9v7r" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.511543 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bt4zh\" (UniqueName: \"kubernetes.io/projected/9586cac6-9d7e-4c04-9400-b0760540e65c-kube-api-access-bt4zh\") pod \"redhat-marketplace-l9v7r\" (UID: \"9586cac6-9d7e-4c04-9400-b0760540e65c\") " pod="openshift-marketplace/redhat-marketplace-l9v7r" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.512243 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9586cac6-9d7e-4c04-9400-b0760540e65c-catalog-content\") pod \"redhat-marketplace-l9v7r\" (UID: \"9586cac6-9d7e-4c04-9400-b0760540e65c\") " pod="openshift-marketplace/redhat-marketplace-l9v7r" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.512510 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9586cac6-9d7e-4c04-9400-b0760540e65c-utilities\") pod \"redhat-marketplace-l9v7r\" (UID: \"9586cac6-9d7e-4c04-9400-b0760540e65c\") " pod="openshift-marketplace/redhat-marketplace-l9v7r" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.578904 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bt4zh\" (UniqueName: \"kubernetes.io/projected/9586cac6-9d7e-4c04-9400-b0760540e65c-kube-api-access-bt4zh\") pod \"redhat-marketplace-l9v7r\" (UID: \"9586cac6-9d7e-4c04-9400-b0760540e65c\") " pod="openshift-marketplace/redhat-marketplace-l9v7r" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.589684 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w5ghj"] Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.590589 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w5ghj" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.592436 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.602410 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w5ghj"] Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.611938 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e09c138-944f-42fa-b52f-cc3cf1395226-utilities\") pod \"redhat-operators-w5ghj\" (UID: \"2e09c138-944f-42fa-b52f-cc3cf1395226\") " pod="openshift-marketplace/redhat-operators-w5ghj" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.612025 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqcq8\" (UniqueName: \"kubernetes.io/projected/2e09c138-944f-42fa-b52f-cc3cf1395226-kube-api-access-tqcq8\") pod \"redhat-operators-w5ghj\" (UID: \"2e09c138-944f-42fa-b52f-cc3cf1395226\") " pod="openshift-marketplace/redhat-operators-w5ghj" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.612066 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e09c138-944f-42fa-b52f-cc3cf1395226-catalog-content\") pod \"redhat-operators-w5ghj\" (UID: \"2e09c138-944f-42fa-b52f-cc3cf1395226\") " pod="openshift-marketplace/redhat-operators-w5ghj" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.680060 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l9v7r" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.712990 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e09c138-944f-42fa-b52f-cc3cf1395226-catalog-content\") pod \"redhat-operators-w5ghj\" (UID: \"2e09c138-944f-42fa-b52f-cc3cf1395226\") " pod="openshift-marketplace/redhat-operators-w5ghj" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.713053 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e09c138-944f-42fa-b52f-cc3cf1395226-utilities\") pod \"redhat-operators-w5ghj\" (UID: \"2e09c138-944f-42fa-b52f-cc3cf1395226\") " pod="openshift-marketplace/redhat-operators-w5ghj" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.713097 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqcq8\" (UniqueName: \"kubernetes.io/projected/2e09c138-944f-42fa-b52f-cc3cf1395226-kube-api-access-tqcq8\") pod \"redhat-operators-w5ghj\" (UID: \"2e09c138-944f-42fa-b52f-cc3cf1395226\") " pod="openshift-marketplace/redhat-operators-w5ghj" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.713807 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e09c138-944f-42fa-b52f-cc3cf1395226-catalog-content\") pod \"redhat-operators-w5ghj\" (UID: \"2e09c138-944f-42fa-b52f-cc3cf1395226\") " pod="openshift-marketplace/redhat-operators-w5ghj" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.713931 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e09c138-944f-42fa-b52f-cc3cf1395226-utilities\") pod \"redhat-operators-w5ghj\" (UID: \"2e09c138-944f-42fa-b52f-cc3cf1395226\") " pod="openshift-marketplace/redhat-operators-w5ghj" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.731713 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqcq8\" (UniqueName: \"kubernetes.io/projected/2e09c138-944f-42fa-b52f-cc3cf1395226-kube-api-access-tqcq8\") pod \"redhat-operators-w5ghj\" (UID: \"2e09c138-944f-42fa-b52f-cc3cf1395226\") " pod="openshift-marketplace/redhat-operators-w5ghj" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.856604 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l9v7r"] Nov 24 06:59:29 crc kubenswrapper[4809]: W1124 06:59:29.868080 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9586cac6_9d7e_4c04_9400_b0760540e65c.slice/crio-3f81f754ac725ff499bc120181039ebc917a65ac46e3c3a8488edc745239513b WatchSource:0}: Error finding container 3f81f754ac725ff499bc120181039ebc917a65ac46e3c3a8488edc745239513b: Status 404 returned error can't find the container with id 3f81f754ac725ff499bc120181039ebc917a65ac46e3c3a8488edc745239513b Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.919254 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w5ghj" Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.992646 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l9v7r" event={"ID":"9586cac6-9d7e-4c04-9400-b0760540e65c","Type":"ContainerStarted","Data":"151adaf0e146e99bc00b22b3918d70917f6311dd881196ffc5c99c3634a0f821"} Nov 24 06:59:29 crc kubenswrapper[4809]: I1124 06:59:29.992682 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l9v7r" event={"ID":"9586cac6-9d7e-4c04-9400-b0760540e65c","Type":"ContainerStarted","Data":"3f81f754ac725ff499bc120181039ebc917a65ac46e3c3a8488edc745239513b"} Nov 24 06:59:30 crc kubenswrapper[4809]: I1124 06:59:30.287692 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w5ghj"] Nov 24 06:59:30 crc kubenswrapper[4809]: W1124 06:59:30.297147 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e09c138_944f_42fa_b52f_cc3cf1395226.slice/crio-d07072d5b359b48c377cf55aa2ce6d55316d1fd5403ee37c7884d3f5817b7690 WatchSource:0}: Error finding container d07072d5b359b48c377cf55aa2ce6d55316d1fd5403ee37c7884d3f5817b7690: Status 404 returned error can't find the container with id d07072d5b359b48c377cf55aa2ce6d55316d1fd5403ee37c7884d3f5817b7690 Nov 24 06:59:31 crc kubenswrapper[4809]: I1124 06:59:30.999907 4809 generic.go:334] "Generic (PLEG): container finished" podID="2e09c138-944f-42fa-b52f-cc3cf1395226" containerID="245926ab3d8a7793dddb426aaa948f234738df85b6238085ecdda67451f0e0bf" exitCode=0 Nov 24 06:59:31 crc kubenswrapper[4809]: I1124 06:59:30.999988 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5ghj" event={"ID":"2e09c138-944f-42fa-b52f-cc3cf1395226","Type":"ContainerDied","Data":"245926ab3d8a7793dddb426aaa948f234738df85b6238085ecdda67451f0e0bf"} Nov 24 06:59:31 crc kubenswrapper[4809]: I1124 06:59:31.000322 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5ghj" event={"ID":"2e09c138-944f-42fa-b52f-cc3cf1395226","Type":"ContainerStarted","Data":"d07072d5b359b48c377cf55aa2ce6d55316d1fd5403ee37c7884d3f5817b7690"} Nov 24 06:59:31 crc kubenswrapper[4809]: I1124 06:59:31.001778 4809 generic.go:334] "Generic (PLEG): container finished" podID="9586cac6-9d7e-4c04-9400-b0760540e65c" containerID="151adaf0e146e99bc00b22b3918d70917f6311dd881196ffc5c99c3634a0f821" exitCode=0 Nov 24 06:59:31 crc kubenswrapper[4809]: I1124 06:59:31.001826 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l9v7r" event={"ID":"9586cac6-9d7e-4c04-9400-b0760540e65c","Type":"ContainerDied","Data":"151adaf0e146e99bc00b22b3918d70917f6311dd881196ffc5c99c3634a0f821"} Nov 24 06:59:31 crc kubenswrapper[4809]: I1124 06:59:31.777840 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7lv57"] Nov 24 06:59:31 crc kubenswrapper[4809]: I1124 06:59:31.781344 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7lv57" Nov 24 06:59:31 crc kubenswrapper[4809]: I1124 06:59:31.785237 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 06:59:31 crc kubenswrapper[4809]: I1124 06:59:31.789268 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7lv57"] Nov 24 06:59:31 crc kubenswrapper[4809]: I1124 06:59:31.834341 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a56840cb-71cf-40df-8df8-c91866948a88-utilities\") pod \"certified-operators-7lv57\" (UID: \"a56840cb-71cf-40df-8df8-c91866948a88\") " pod="openshift-marketplace/certified-operators-7lv57" Nov 24 06:59:31 crc kubenswrapper[4809]: I1124 06:59:31.834385 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a56840cb-71cf-40df-8df8-c91866948a88-catalog-content\") pod \"certified-operators-7lv57\" (UID: \"a56840cb-71cf-40df-8df8-c91866948a88\") " pod="openshift-marketplace/certified-operators-7lv57" Nov 24 06:59:31 crc kubenswrapper[4809]: I1124 06:59:31.834558 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg9fq\" (UniqueName: \"kubernetes.io/projected/a56840cb-71cf-40df-8df8-c91866948a88-kube-api-access-jg9fq\") pod \"certified-operators-7lv57\" (UID: \"a56840cb-71cf-40df-8df8-c91866948a88\") " pod="openshift-marketplace/certified-operators-7lv57" Nov 24 06:59:31 crc kubenswrapper[4809]: I1124 06:59:31.935101 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg9fq\" (UniqueName: \"kubernetes.io/projected/a56840cb-71cf-40df-8df8-c91866948a88-kube-api-access-jg9fq\") pod \"certified-operators-7lv57\" (UID: \"a56840cb-71cf-40df-8df8-c91866948a88\") " pod="openshift-marketplace/certified-operators-7lv57" Nov 24 06:59:31 crc kubenswrapper[4809]: I1124 06:59:31.935165 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a56840cb-71cf-40df-8df8-c91866948a88-utilities\") pod \"certified-operators-7lv57\" (UID: \"a56840cb-71cf-40df-8df8-c91866948a88\") " pod="openshift-marketplace/certified-operators-7lv57" Nov 24 06:59:31 crc kubenswrapper[4809]: I1124 06:59:31.935196 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a56840cb-71cf-40df-8df8-c91866948a88-catalog-content\") pod \"certified-operators-7lv57\" (UID: \"a56840cb-71cf-40df-8df8-c91866948a88\") " pod="openshift-marketplace/certified-operators-7lv57" Nov 24 06:59:31 crc kubenswrapper[4809]: I1124 06:59:31.935627 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a56840cb-71cf-40df-8df8-c91866948a88-catalog-content\") pod \"certified-operators-7lv57\" (UID: \"a56840cb-71cf-40df-8df8-c91866948a88\") " pod="openshift-marketplace/certified-operators-7lv57" Nov 24 06:59:31 crc kubenswrapper[4809]: I1124 06:59:31.935952 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a56840cb-71cf-40df-8df8-c91866948a88-utilities\") pod \"certified-operators-7lv57\" (UID: \"a56840cb-71cf-40df-8df8-c91866948a88\") " pod="openshift-marketplace/certified-operators-7lv57" Nov 24 06:59:31 crc kubenswrapper[4809]: I1124 06:59:31.959381 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg9fq\" (UniqueName: \"kubernetes.io/projected/a56840cb-71cf-40df-8df8-c91866948a88-kube-api-access-jg9fq\") pod \"certified-operators-7lv57\" (UID: \"a56840cb-71cf-40df-8df8-c91866948a88\") " pod="openshift-marketplace/certified-operators-7lv57" Nov 24 06:59:31 crc kubenswrapper[4809]: I1124 06:59:31.986276 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rhg7d"] Nov 24 06:59:31 crc kubenswrapper[4809]: I1124 06:59:31.987462 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rhg7d" Nov 24 06:59:31 crc kubenswrapper[4809]: I1124 06:59:31.989889 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 06:59:31 crc kubenswrapper[4809]: I1124 06:59:31.992278 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rhg7d"] Nov 24 06:59:32 crc kubenswrapper[4809]: I1124 06:59:32.011223 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5ghj" event={"ID":"2e09c138-944f-42fa-b52f-cc3cf1395226","Type":"ContainerStarted","Data":"0264b58e2db49abf24a5d4368e0c97913c433a9e81226f9b7e1b742d988d08cd"} Nov 24 06:59:32 crc kubenswrapper[4809]: I1124 06:59:32.035889 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41964283-35a3-4b8d-8e54-c3c7518f62ee-catalog-content\") pod \"community-operators-rhg7d\" (UID: \"41964283-35a3-4b8d-8e54-c3c7518f62ee\") " pod="openshift-marketplace/community-operators-rhg7d" Nov 24 06:59:32 crc kubenswrapper[4809]: I1124 06:59:32.036203 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nnb5\" (UniqueName: \"kubernetes.io/projected/41964283-35a3-4b8d-8e54-c3c7518f62ee-kube-api-access-6nnb5\") pod \"community-operators-rhg7d\" (UID: \"41964283-35a3-4b8d-8e54-c3c7518f62ee\") " pod="openshift-marketplace/community-operators-rhg7d" Nov 24 06:59:32 crc kubenswrapper[4809]: I1124 06:59:32.036326 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41964283-35a3-4b8d-8e54-c3c7518f62ee-utilities\") pod \"community-operators-rhg7d\" (UID: \"41964283-35a3-4b8d-8e54-c3c7518f62ee\") " pod="openshift-marketplace/community-operators-rhg7d" Nov 24 06:59:32 crc kubenswrapper[4809]: I1124 06:59:32.133614 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7lv57" Nov 24 06:59:32 crc kubenswrapper[4809]: I1124 06:59:32.143440 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41964283-35a3-4b8d-8e54-c3c7518f62ee-catalog-content\") pod \"community-operators-rhg7d\" (UID: \"41964283-35a3-4b8d-8e54-c3c7518f62ee\") " pod="openshift-marketplace/community-operators-rhg7d" Nov 24 06:59:32 crc kubenswrapper[4809]: I1124 06:59:32.143495 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nnb5\" (UniqueName: \"kubernetes.io/projected/41964283-35a3-4b8d-8e54-c3c7518f62ee-kube-api-access-6nnb5\") pod \"community-operators-rhg7d\" (UID: \"41964283-35a3-4b8d-8e54-c3c7518f62ee\") " pod="openshift-marketplace/community-operators-rhg7d" Nov 24 06:59:32 crc kubenswrapper[4809]: I1124 06:59:32.143528 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41964283-35a3-4b8d-8e54-c3c7518f62ee-utilities\") pod \"community-operators-rhg7d\" (UID: \"41964283-35a3-4b8d-8e54-c3c7518f62ee\") " pod="openshift-marketplace/community-operators-rhg7d" Nov 24 06:59:32 crc kubenswrapper[4809]: I1124 06:59:32.144167 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41964283-35a3-4b8d-8e54-c3c7518f62ee-catalog-content\") pod \"community-operators-rhg7d\" (UID: \"41964283-35a3-4b8d-8e54-c3c7518f62ee\") " pod="openshift-marketplace/community-operators-rhg7d" Nov 24 06:59:32 crc kubenswrapper[4809]: I1124 06:59:32.144894 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41964283-35a3-4b8d-8e54-c3c7518f62ee-utilities\") pod \"community-operators-rhg7d\" (UID: \"41964283-35a3-4b8d-8e54-c3c7518f62ee\") " pod="openshift-marketplace/community-operators-rhg7d" Nov 24 06:59:32 crc kubenswrapper[4809]: I1124 06:59:32.170102 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nnb5\" (UniqueName: \"kubernetes.io/projected/41964283-35a3-4b8d-8e54-c3c7518f62ee-kube-api-access-6nnb5\") pod \"community-operators-rhg7d\" (UID: \"41964283-35a3-4b8d-8e54-c3c7518f62ee\") " pod="openshift-marketplace/community-operators-rhg7d" Nov 24 06:59:32 crc kubenswrapper[4809]: I1124 06:59:32.358512 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rhg7d" Nov 24 06:59:32 crc kubenswrapper[4809]: I1124 06:59:32.525616 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7lv57"] Nov 24 06:59:32 crc kubenswrapper[4809]: W1124 06:59:32.531034 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda56840cb_71cf_40df_8df8_c91866948a88.slice/crio-4a9d89e9f126ceb3605c8b448a28bd7134155720d5a8a86a7762b21451f23755 WatchSource:0}: Error finding container 4a9d89e9f126ceb3605c8b448a28bd7134155720d5a8a86a7762b21451f23755: Status 404 returned error can't find the container with id 4a9d89e9f126ceb3605c8b448a28bd7134155720d5a8a86a7762b21451f23755 Nov 24 06:59:32 crc kubenswrapper[4809]: I1124 06:59:32.724030 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rhg7d"] Nov 24 06:59:32 crc kubenswrapper[4809]: W1124 06:59:32.738330 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod41964283_35a3_4b8d_8e54_c3c7518f62ee.slice/crio-164aca2d37836075cf656b2386a2d8d275297ccb2a1b86baef840d7e2cfc09ee WatchSource:0}: Error finding container 164aca2d37836075cf656b2386a2d8d275297ccb2a1b86baef840d7e2cfc09ee: Status 404 returned error can't find the container with id 164aca2d37836075cf656b2386a2d8d275297ccb2a1b86baef840d7e2cfc09ee Nov 24 06:59:33 crc kubenswrapper[4809]: I1124 06:59:33.016459 4809 generic.go:334] "Generic (PLEG): container finished" podID="a56840cb-71cf-40df-8df8-c91866948a88" containerID="02d5564f6f7488975a4f659d5360d1446a75cd0b74933cab0bccddb17a0da8e4" exitCode=0 Nov 24 06:59:33 crc kubenswrapper[4809]: I1124 06:59:33.016651 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7lv57" event={"ID":"a56840cb-71cf-40df-8df8-c91866948a88","Type":"ContainerDied","Data":"02d5564f6f7488975a4f659d5360d1446a75cd0b74933cab0bccddb17a0da8e4"} Nov 24 06:59:33 crc kubenswrapper[4809]: I1124 06:59:33.016754 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7lv57" event={"ID":"a56840cb-71cf-40df-8df8-c91866948a88","Type":"ContainerStarted","Data":"4a9d89e9f126ceb3605c8b448a28bd7134155720d5a8a86a7762b21451f23755"} Nov 24 06:59:33 crc kubenswrapper[4809]: I1124 06:59:33.018466 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5ghj" event={"ID":"2e09c138-944f-42fa-b52f-cc3cf1395226","Type":"ContainerDied","Data":"0264b58e2db49abf24a5d4368e0c97913c433a9e81226f9b7e1b742d988d08cd"} Nov 24 06:59:33 crc kubenswrapper[4809]: I1124 06:59:33.018306 4809 generic.go:334] "Generic (PLEG): container finished" podID="2e09c138-944f-42fa-b52f-cc3cf1395226" containerID="0264b58e2db49abf24a5d4368e0c97913c433a9e81226f9b7e1b742d988d08cd" exitCode=0 Nov 24 06:59:33 crc kubenswrapper[4809]: I1124 06:59:33.022589 4809 generic.go:334] "Generic (PLEG): container finished" podID="9586cac6-9d7e-4c04-9400-b0760540e65c" containerID="3d1aa3653ca1de15f96859a2cc4685d5972a5f46eaac62cfdb11926f5ef9fe6c" exitCode=0 Nov 24 06:59:33 crc kubenswrapper[4809]: I1124 06:59:33.022884 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l9v7r" event={"ID":"9586cac6-9d7e-4c04-9400-b0760540e65c","Type":"ContainerDied","Data":"3d1aa3653ca1de15f96859a2cc4685d5972a5f46eaac62cfdb11926f5ef9fe6c"} Nov 24 06:59:33 crc kubenswrapper[4809]: I1124 06:59:33.033375 4809 generic.go:334] "Generic (PLEG): container finished" podID="41964283-35a3-4b8d-8e54-c3c7518f62ee" containerID="03130db63b837c63ae279deb3214140c028d66ff03bbd571f524d761c68f8148" exitCode=0 Nov 24 06:59:33 crc kubenswrapper[4809]: I1124 06:59:33.033410 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rhg7d" event={"ID":"41964283-35a3-4b8d-8e54-c3c7518f62ee","Type":"ContainerDied","Data":"03130db63b837c63ae279deb3214140c028d66ff03bbd571f524d761c68f8148"} Nov 24 06:59:33 crc kubenswrapper[4809]: I1124 06:59:33.033433 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rhg7d" event={"ID":"41964283-35a3-4b8d-8e54-c3c7518f62ee","Type":"ContainerStarted","Data":"164aca2d37836075cf656b2386a2d8d275297ccb2a1b86baef840d7e2cfc09ee"} Nov 24 06:59:34 crc kubenswrapper[4809]: I1124 06:59:34.040035 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l9v7r" event={"ID":"9586cac6-9d7e-4c04-9400-b0760540e65c","Type":"ContainerStarted","Data":"fe428241ea09a5a776de80874f891c9ea2a709851cb7ee2f269b9bf90175f970"} Nov 24 06:59:34 crc kubenswrapper[4809]: I1124 06:59:34.042701 4809 generic.go:334] "Generic (PLEG): container finished" podID="41964283-35a3-4b8d-8e54-c3c7518f62ee" containerID="15bb3bd9449fd67416a8a8e56167b5b38291595472c72e75431227d385a2bf27" exitCode=0 Nov 24 06:59:34 crc kubenswrapper[4809]: I1124 06:59:34.042753 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rhg7d" event={"ID":"41964283-35a3-4b8d-8e54-c3c7518f62ee","Type":"ContainerDied","Data":"15bb3bd9449fd67416a8a8e56167b5b38291595472c72e75431227d385a2bf27"} Nov 24 06:59:34 crc kubenswrapper[4809]: I1124 06:59:34.045587 4809 generic.go:334] "Generic (PLEG): container finished" podID="a56840cb-71cf-40df-8df8-c91866948a88" containerID="721caff5f151f85cb069840ecf58461a1f29a4c43f728cff4f5f4be3de52868c" exitCode=0 Nov 24 06:59:34 crc kubenswrapper[4809]: I1124 06:59:34.045625 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7lv57" event={"ID":"a56840cb-71cf-40df-8df8-c91866948a88","Type":"ContainerDied","Data":"721caff5f151f85cb069840ecf58461a1f29a4c43f728cff4f5f4be3de52868c"} Nov 24 06:59:34 crc kubenswrapper[4809]: I1124 06:59:34.049321 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5ghj" event={"ID":"2e09c138-944f-42fa-b52f-cc3cf1395226","Type":"ContainerStarted","Data":"0737d3bdf4f5ef9427ec877a4078201a87fc15cc6ea2b40027b831bea5805efe"} Nov 24 06:59:34 crc kubenswrapper[4809]: I1124 06:59:34.066368 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-l9v7r" podStartSLOduration=2.659911614 podStartE2EDuration="5.066351128s" podCreationTimestamp="2025-11-24 06:59:29 +0000 UTC" firstStartedPulling="2025-11-24 06:59:31.003749793 +0000 UTC m=+270.904341498" lastFinishedPulling="2025-11-24 06:59:33.410189307 +0000 UTC m=+273.310781012" observedRunningTime="2025-11-24 06:59:34.064028391 +0000 UTC m=+273.964620086" watchObservedRunningTime="2025-11-24 06:59:34.066351128 +0000 UTC m=+273.966942833" Nov 24 06:59:34 crc kubenswrapper[4809]: I1124 06:59:34.113539 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w5ghj" podStartSLOduration=2.658830719 podStartE2EDuration="5.11350625s" podCreationTimestamp="2025-11-24 06:59:29 +0000 UTC" firstStartedPulling="2025-11-24 06:59:31.001602162 +0000 UTC m=+270.902193887" lastFinishedPulling="2025-11-24 06:59:33.456277713 +0000 UTC m=+273.356869418" observedRunningTime="2025-11-24 06:59:34.111842775 +0000 UTC m=+274.012434490" watchObservedRunningTime="2025-11-24 06:59:34.11350625 +0000 UTC m=+274.014097955" Nov 24 06:59:35 crc kubenswrapper[4809]: I1124 06:59:35.055233 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7lv57" event={"ID":"a56840cb-71cf-40df-8df8-c91866948a88","Type":"ContainerStarted","Data":"0aee0ae75181087270df3bada0e8ce0c557798147ac9d770d04148fc3218629b"} Nov 24 06:59:35 crc kubenswrapper[4809]: I1124 06:59:35.057173 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rhg7d" event={"ID":"41964283-35a3-4b8d-8e54-c3c7518f62ee","Type":"ContainerStarted","Data":"ff3d073588c45d12604f414b860d30c70c95f29b7bfc326e4990a89750ef5d91"} Nov 24 06:59:35 crc kubenswrapper[4809]: I1124 06:59:35.073574 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7lv57" podStartSLOduration=2.662024574 podStartE2EDuration="4.073553652s" podCreationTimestamp="2025-11-24 06:59:31 +0000 UTC" firstStartedPulling="2025-11-24 06:59:33.019078969 +0000 UTC m=+272.919670684" lastFinishedPulling="2025-11-24 06:59:34.430608057 +0000 UTC m=+274.331199762" observedRunningTime="2025-11-24 06:59:35.070513321 +0000 UTC m=+274.971105026" watchObservedRunningTime="2025-11-24 06:59:35.073553652 +0000 UTC m=+274.974145357" Nov 24 06:59:39 crc kubenswrapper[4809]: I1124 06:59:39.683242 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-l9v7r" Nov 24 06:59:39 crc kubenswrapper[4809]: I1124 06:59:39.684795 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-l9v7r" Nov 24 06:59:39 crc kubenswrapper[4809]: I1124 06:59:39.727590 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-l9v7r" Nov 24 06:59:39 crc kubenswrapper[4809]: I1124 06:59:39.742063 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rhg7d" podStartSLOduration=7.281417905 podStartE2EDuration="8.742044649s" podCreationTimestamp="2025-11-24 06:59:31 +0000 UTC" firstStartedPulling="2025-11-24 06:59:33.034919534 +0000 UTC m=+272.935511239" lastFinishedPulling="2025-11-24 06:59:34.495546278 +0000 UTC m=+274.396137983" observedRunningTime="2025-11-24 06:59:35.092478219 +0000 UTC m=+274.993069924" watchObservedRunningTime="2025-11-24 06:59:39.742044649 +0000 UTC m=+279.642636364" Nov 24 06:59:39 crc kubenswrapper[4809]: I1124 06:59:39.919876 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w5ghj" Nov 24 06:59:39 crc kubenswrapper[4809]: I1124 06:59:39.919995 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w5ghj" Nov 24 06:59:39 crc kubenswrapper[4809]: I1124 06:59:39.969252 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w5ghj" Nov 24 06:59:40 crc kubenswrapper[4809]: I1124 06:59:40.121496 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-l9v7r" Nov 24 06:59:40 crc kubenswrapper[4809]: I1124 06:59:40.122696 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w5ghj" Nov 24 06:59:42 crc kubenswrapper[4809]: I1124 06:59:42.134368 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7lv57" Nov 24 06:59:42 crc kubenswrapper[4809]: I1124 06:59:42.134851 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7lv57" Nov 24 06:59:42 crc kubenswrapper[4809]: I1124 06:59:42.174774 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7lv57" Nov 24 06:59:42 crc kubenswrapper[4809]: I1124 06:59:42.359348 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rhg7d" Nov 24 06:59:42 crc kubenswrapper[4809]: I1124 06:59:42.359743 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rhg7d" Nov 24 06:59:42 crc kubenswrapper[4809]: I1124 06:59:42.395417 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rhg7d" Nov 24 06:59:43 crc kubenswrapper[4809]: I1124 06:59:43.155411 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rhg7d" Nov 24 06:59:43 crc kubenswrapper[4809]: I1124 06:59:43.156501 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7lv57" Nov 24 07:00:00 crc kubenswrapper[4809]: I1124 07:00:00.142609 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399460-hdrql"] Nov 24 07:00:00 crc kubenswrapper[4809]: I1124 07:00:00.144874 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-hdrql" Nov 24 07:00:00 crc kubenswrapper[4809]: I1124 07:00:00.146892 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 07:00:00 crc kubenswrapper[4809]: I1124 07:00:00.147207 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 07:00:00 crc kubenswrapper[4809]: I1124 07:00:00.147221 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399460-hdrql"] Nov 24 07:00:00 crc kubenswrapper[4809]: I1124 07:00:00.295432 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghvh8\" (UniqueName: \"kubernetes.io/projected/2f2df143-a48b-46eb-9b59-3903504fe5a4-kube-api-access-ghvh8\") pod \"collect-profiles-29399460-hdrql\" (UID: \"2f2df143-a48b-46eb-9b59-3903504fe5a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-hdrql" Nov 24 07:00:00 crc kubenswrapper[4809]: I1124 07:00:00.295489 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f2df143-a48b-46eb-9b59-3903504fe5a4-config-volume\") pod \"collect-profiles-29399460-hdrql\" (UID: \"2f2df143-a48b-46eb-9b59-3903504fe5a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-hdrql" Nov 24 07:00:00 crc kubenswrapper[4809]: I1124 07:00:00.295723 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f2df143-a48b-46eb-9b59-3903504fe5a4-secret-volume\") pod \"collect-profiles-29399460-hdrql\" (UID: \"2f2df143-a48b-46eb-9b59-3903504fe5a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-hdrql" Nov 24 07:00:00 crc kubenswrapper[4809]: I1124 07:00:00.397102 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f2df143-a48b-46eb-9b59-3903504fe5a4-secret-volume\") pod \"collect-profiles-29399460-hdrql\" (UID: \"2f2df143-a48b-46eb-9b59-3903504fe5a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-hdrql" Nov 24 07:00:00 crc kubenswrapper[4809]: I1124 07:00:00.397198 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghvh8\" (UniqueName: \"kubernetes.io/projected/2f2df143-a48b-46eb-9b59-3903504fe5a4-kube-api-access-ghvh8\") pod \"collect-profiles-29399460-hdrql\" (UID: \"2f2df143-a48b-46eb-9b59-3903504fe5a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-hdrql" Nov 24 07:00:00 crc kubenswrapper[4809]: I1124 07:00:00.397219 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f2df143-a48b-46eb-9b59-3903504fe5a4-config-volume\") pod \"collect-profiles-29399460-hdrql\" (UID: \"2f2df143-a48b-46eb-9b59-3903504fe5a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-hdrql" Nov 24 07:00:00 crc kubenswrapper[4809]: I1124 07:00:00.398321 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f2df143-a48b-46eb-9b59-3903504fe5a4-config-volume\") pod \"collect-profiles-29399460-hdrql\" (UID: \"2f2df143-a48b-46eb-9b59-3903504fe5a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-hdrql" Nov 24 07:00:00 crc kubenswrapper[4809]: I1124 07:00:00.402233 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f2df143-a48b-46eb-9b59-3903504fe5a4-secret-volume\") pod \"collect-profiles-29399460-hdrql\" (UID: \"2f2df143-a48b-46eb-9b59-3903504fe5a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-hdrql" Nov 24 07:00:00 crc kubenswrapper[4809]: I1124 07:00:00.418161 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghvh8\" (UniqueName: \"kubernetes.io/projected/2f2df143-a48b-46eb-9b59-3903504fe5a4-kube-api-access-ghvh8\") pod \"collect-profiles-29399460-hdrql\" (UID: \"2f2df143-a48b-46eb-9b59-3903504fe5a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-hdrql" Nov 24 07:00:00 crc kubenswrapper[4809]: I1124 07:00:00.468455 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-hdrql" Nov 24 07:00:00 crc kubenswrapper[4809]: I1124 07:00:00.642078 4809 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Nov 24 07:00:00 crc kubenswrapper[4809]: I1124 07:00:00.663452 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399460-hdrql"] Nov 24 07:00:01 crc kubenswrapper[4809]: I1124 07:00:01.186438 4809 generic.go:334] "Generic (PLEG): container finished" podID="2f2df143-a48b-46eb-9b59-3903504fe5a4" containerID="636cfddffffee7ce696a864d7631a31103588c2f5d13834a00a576d89f6e95b4" exitCode=0 Nov 24 07:00:01 crc kubenswrapper[4809]: I1124 07:00:01.186497 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-hdrql" event={"ID":"2f2df143-a48b-46eb-9b59-3903504fe5a4","Type":"ContainerDied","Data":"636cfddffffee7ce696a864d7631a31103588c2f5d13834a00a576d89f6e95b4"} Nov 24 07:00:01 crc kubenswrapper[4809]: I1124 07:00:01.186757 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-hdrql" event={"ID":"2f2df143-a48b-46eb-9b59-3903504fe5a4","Type":"ContainerStarted","Data":"0c5dad591f1422eb62b5d77b8a579f8adb78686b244fd7d5beb9a6a7cd4a5cb2"} Nov 24 07:00:02 crc kubenswrapper[4809]: I1124 07:00:02.452575 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-hdrql" Nov 24 07:00:02 crc kubenswrapper[4809]: I1124 07:00:02.523405 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f2df143-a48b-46eb-9b59-3903504fe5a4-secret-volume\") pod \"2f2df143-a48b-46eb-9b59-3903504fe5a4\" (UID: \"2f2df143-a48b-46eb-9b59-3903504fe5a4\") " Nov 24 07:00:02 crc kubenswrapper[4809]: I1124 07:00:02.524162 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f2df143-a48b-46eb-9b59-3903504fe5a4-config-volume\") pod \"2f2df143-a48b-46eb-9b59-3903504fe5a4\" (UID: \"2f2df143-a48b-46eb-9b59-3903504fe5a4\") " Nov 24 07:00:02 crc kubenswrapper[4809]: I1124 07:00:02.524312 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghvh8\" (UniqueName: \"kubernetes.io/projected/2f2df143-a48b-46eb-9b59-3903504fe5a4-kube-api-access-ghvh8\") pod \"2f2df143-a48b-46eb-9b59-3903504fe5a4\" (UID: \"2f2df143-a48b-46eb-9b59-3903504fe5a4\") " Nov 24 07:00:02 crc kubenswrapper[4809]: I1124 07:00:02.524928 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f2df143-a48b-46eb-9b59-3903504fe5a4-config-volume" (OuterVolumeSpecName: "config-volume") pod "2f2df143-a48b-46eb-9b59-3903504fe5a4" (UID: "2f2df143-a48b-46eb-9b59-3903504fe5a4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:00:02 crc kubenswrapper[4809]: I1124 07:00:02.528493 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f2df143-a48b-46eb-9b59-3903504fe5a4-kube-api-access-ghvh8" (OuterVolumeSpecName: "kube-api-access-ghvh8") pod "2f2df143-a48b-46eb-9b59-3903504fe5a4" (UID: "2f2df143-a48b-46eb-9b59-3903504fe5a4"). InnerVolumeSpecName "kube-api-access-ghvh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:00:02 crc kubenswrapper[4809]: I1124 07:00:02.528608 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f2df143-a48b-46eb-9b59-3903504fe5a4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2f2df143-a48b-46eb-9b59-3903504fe5a4" (UID: "2f2df143-a48b-46eb-9b59-3903504fe5a4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:00:02 crc kubenswrapper[4809]: I1124 07:00:02.625453 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f2df143-a48b-46eb-9b59-3903504fe5a4-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 07:00:02 crc kubenswrapper[4809]: I1124 07:00:02.625485 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f2df143-a48b-46eb-9b59-3903504fe5a4-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 07:00:02 crc kubenswrapper[4809]: I1124 07:00:02.625497 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghvh8\" (UniqueName: \"kubernetes.io/projected/2f2df143-a48b-46eb-9b59-3903504fe5a4-kube-api-access-ghvh8\") on node \"crc\" DevicePath \"\"" Nov 24 07:00:03 crc kubenswrapper[4809]: I1124 07:00:03.199345 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-hdrql" event={"ID":"2f2df143-a48b-46eb-9b59-3903504fe5a4","Type":"ContainerDied","Data":"0c5dad591f1422eb62b5d77b8a579f8adb78686b244fd7d5beb9a6a7cd4a5cb2"} Nov 24 07:00:03 crc kubenswrapper[4809]: I1124 07:00:03.199384 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c5dad591f1422eb62b5d77b8a579f8adb78686b244fd7d5beb9a6a7cd4a5cb2" Nov 24 07:00:03 crc kubenswrapper[4809]: I1124 07:00:03.199411 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-hdrql" Nov 24 07:00:48 crc kubenswrapper[4809]: I1124 07:00:48.043571 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:00:48 crc kubenswrapper[4809]: I1124 07:00:48.044379 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:01:18 crc kubenswrapper[4809]: I1124 07:01:18.043463 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:01:18 crc kubenswrapper[4809]: I1124 07:01:18.044106 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:01:48 crc kubenswrapper[4809]: I1124 07:01:48.043616 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:01:48 crc kubenswrapper[4809]: I1124 07:01:48.045142 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:01:48 crc kubenswrapper[4809]: I1124 07:01:48.045237 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 07:01:48 crc kubenswrapper[4809]: I1124 07:01:48.046472 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"33316ec1d7739e9081fef97a1f1a3a1c2f00b3bbab2a32058a1c6e7ff1f92aad"} pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:01:48 crc kubenswrapper[4809]: I1124 07:01:48.046601 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" containerID="cri-o://33316ec1d7739e9081fef97a1f1a3a1c2f00b3bbab2a32058a1c6e7ff1f92aad" gracePeriod=600 Nov 24 07:01:48 crc kubenswrapper[4809]: I1124 07:01:48.786240 4809 generic.go:334] "Generic (PLEG): container finished" podID="be8050d5-1aea-453e-a06e-87702b0e856b" containerID="33316ec1d7739e9081fef97a1f1a3a1c2f00b3bbab2a32058a1c6e7ff1f92aad" exitCode=0 Nov 24 07:01:48 crc kubenswrapper[4809]: I1124 07:01:48.786316 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerDied","Data":"33316ec1d7739e9081fef97a1f1a3a1c2f00b3bbab2a32058a1c6e7ff1f92aad"} Nov 24 07:01:48 crc kubenswrapper[4809]: I1124 07:01:48.786589 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerStarted","Data":"24354d43179411329fa3f10b3bcaad9ac5fbb3505999caa72b922a82781a0f78"} Nov 24 07:01:48 crc kubenswrapper[4809]: I1124 07:01:48.786630 4809 scope.go:117] "RemoveContainer" containerID="a00e2a2c556bbc3b9c8d47e36df5c57ead2da64726c51125a67d372a4e4f51d5" Nov 24 07:02:29 crc kubenswrapper[4809]: I1124 07:02:29.989680 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bvzqh"] Nov 24 07:02:29 crc kubenswrapper[4809]: E1124 07:02:29.990341 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f2df143-a48b-46eb-9b59-3903504fe5a4" containerName="collect-profiles" Nov 24 07:02:29 crc kubenswrapper[4809]: I1124 07:02:29.990417 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f2df143-a48b-46eb-9b59-3903504fe5a4" containerName="collect-profiles" Nov 24 07:02:29 crc kubenswrapper[4809]: I1124 07:02:29.990517 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f2df143-a48b-46eb-9b59-3903504fe5a4" containerName="collect-profiles" Nov 24 07:02:29 crc kubenswrapper[4809]: I1124 07:02:29.990930 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.012627 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bvzqh"] Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.104154 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3d751728-7cf1-47cc-b6ab-a6207508c6fb-registry-certificates\") pod \"image-registry-66df7c8f76-bvzqh\" (UID: \"3d751728-7cf1-47cc-b6ab-a6207508c6fb\") " pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.104279 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3d751728-7cf1-47cc-b6ab-a6207508c6fb-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bvzqh\" (UID: \"3d751728-7cf1-47cc-b6ab-a6207508c6fb\") " pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.104309 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3d751728-7cf1-47cc-b6ab-a6207508c6fb-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bvzqh\" (UID: \"3d751728-7cf1-47cc-b6ab-a6207508c6fb\") " pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.104367 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c22m6\" (UniqueName: \"kubernetes.io/projected/3d751728-7cf1-47cc-b6ab-a6207508c6fb-kube-api-access-c22m6\") pod \"image-registry-66df7c8f76-bvzqh\" (UID: \"3d751728-7cf1-47cc-b6ab-a6207508c6fb\") " pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.104403 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3d751728-7cf1-47cc-b6ab-a6207508c6fb-trusted-ca\") pod \"image-registry-66df7c8f76-bvzqh\" (UID: \"3d751728-7cf1-47cc-b6ab-a6207508c6fb\") " pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.104428 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3d751728-7cf1-47cc-b6ab-a6207508c6fb-registry-tls\") pod \"image-registry-66df7c8f76-bvzqh\" (UID: \"3d751728-7cf1-47cc-b6ab-a6207508c6fb\") " pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.104506 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3d751728-7cf1-47cc-b6ab-a6207508c6fb-bound-sa-token\") pod \"image-registry-66df7c8f76-bvzqh\" (UID: \"3d751728-7cf1-47cc-b6ab-a6207508c6fb\") " pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.104809 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bvzqh\" (UID: \"3d751728-7cf1-47cc-b6ab-a6207508c6fb\") " pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.128386 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bvzqh\" (UID: \"3d751728-7cf1-47cc-b6ab-a6207508c6fb\") " pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.208400 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3d751728-7cf1-47cc-b6ab-a6207508c6fb-bound-sa-token\") pod \"image-registry-66df7c8f76-bvzqh\" (UID: \"3d751728-7cf1-47cc-b6ab-a6207508c6fb\") " pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.208542 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3d751728-7cf1-47cc-b6ab-a6207508c6fb-registry-certificates\") pod \"image-registry-66df7c8f76-bvzqh\" (UID: \"3d751728-7cf1-47cc-b6ab-a6207508c6fb\") " pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.208581 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3d751728-7cf1-47cc-b6ab-a6207508c6fb-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bvzqh\" (UID: \"3d751728-7cf1-47cc-b6ab-a6207508c6fb\") " pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.208635 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3d751728-7cf1-47cc-b6ab-a6207508c6fb-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bvzqh\" (UID: \"3d751728-7cf1-47cc-b6ab-a6207508c6fb\") " pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.208672 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c22m6\" (UniqueName: \"kubernetes.io/projected/3d751728-7cf1-47cc-b6ab-a6207508c6fb-kube-api-access-c22m6\") pod \"image-registry-66df7c8f76-bvzqh\" (UID: \"3d751728-7cf1-47cc-b6ab-a6207508c6fb\") " pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.208708 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3d751728-7cf1-47cc-b6ab-a6207508c6fb-registry-tls\") pod \"image-registry-66df7c8f76-bvzqh\" (UID: \"3d751728-7cf1-47cc-b6ab-a6207508c6fb\") " pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.208735 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3d751728-7cf1-47cc-b6ab-a6207508c6fb-trusted-ca\") pod \"image-registry-66df7c8f76-bvzqh\" (UID: \"3d751728-7cf1-47cc-b6ab-a6207508c6fb\") " pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.209308 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3d751728-7cf1-47cc-b6ab-a6207508c6fb-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bvzqh\" (UID: \"3d751728-7cf1-47cc-b6ab-a6207508c6fb\") " pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.210609 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3d751728-7cf1-47cc-b6ab-a6207508c6fb-trusted-ca\") pod \"image-registry-66df7c8f76-bvzqh\" (UID: \"3d751728-7cf1-47cc-b6ab-a6207508c6fb\") " pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.210629 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3d751728-7cf1-47cc-b6ab-a6207508c6fb-registry-certificates\") pod \"image-registry-66df7c8f76-bvzqh\" (UID: \"3d751728-7cf1-47cc-b6ab-a6207508c6fb\") " pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.216464 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3d751728-7cf1-47cc-b6ab-a6207508c6fb-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bvzqh\" (UID: \"3d751728-7cf1-47cc-b6ab-a6207508c6fb\") " pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.218809 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3d751728-7cf1-47cc-b6ab-a6207508c6fb-registry-tls\") pod \"image-registry-66df7c8f76-bvzqh\" (UID: \"3d751728-7cf1-47cc-b6ab-a6207508c6fb\") " pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.232177 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c22m6\" (UniqueName: \"kubernetes.io/projected/3d751728-7cf1-47cc-b6ab-a6207508c6fb-kube-api-access-c22m6\") pod \"image-registry-66df7c8f76-bvzqh\" (UID: \"3d751728-7cf1-47cc-b6ab-a6207508c6fb\") " pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.238595 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3d751728-7cf1-47cc-b6ab-a6207508c6fb-bound-sa-token\") pod \"image-registry-66df7c8f76-bvzqh\" (UID: \"3d751728-7cf1-47cc-b6ab-a6207508c6fb\") " pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.307482 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:30 crc kubenswrapper[4809]: I1124 07:02:30.512236 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bvzqh"] Nov 24 07:02:31 crc kubenswrapper[4809]: I1124 07:02:31.025232 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" event={"ID":"3d751728-7cf1-47cc-b6ab-a6207508c6fb","Type":"ContainerStarted","Data":"9751fd69e1e60778dfafe38baef4ed1e27d37438fb207efb7cbaefc3dfe976b4"} Nov 24 07:02:31 crc kubenswrapper[4809]: I1124 07:02:31.025282 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" event={"ID":"3d751728-7cf1-47cc-b6ab-a6207508c6fb","Type":"ContainerStarted","Data":"6c59b38201d163652034f317370a794f751d2e499dd298a5b0f03ce50ce52cd0"} Nov 24 07:02:31 crc kubenswrapper[4809]: I1124 07:02:31.025382 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:31 crc kubenswrapper[4809]: I1124 07:02:31.044704 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" podStartSLOduration=2.044685199 podStartE2EDuration="2.044685199s" podCreationTimestamp="2025-11-24 07:02:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:02:31.043066895 +0000 UTC m=+450.943658600" watchObservedRunningTime="2025-11-24 07:02:31.044685199 +0000 UTC m=+450.945276904" Nov 24 07:02:50 crc kubenswrapper[4809]: I1124 07:02:50.316009 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-bvzqh" Nov 24 07:02:50 crc kubenswrapper[4809]: I1124 07:02:50.389391 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-59cgv"] Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.422531 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" podUID="4af1418f-a56e-4f26-a800-3b402b7da053" containerName="registry" containerID="cri-o://956426f3fe1032f711950ed48c6b6cff0626cabfb6e09ab8d7118943ffdfcd9e" gracePeriod=30 Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.762101 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.847520 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4af1418f-a56e-4f26-a800-3b402b7da053-installation-pull-secrets\") pod \"4af1418f-a56e-4f26-a800-3b402b7da053\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.847736 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"4af1418f-a56e-4f26-a800-3b402b7da053\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.847790 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4af1418f-a56e-4f26-a800-3b402b7da053-bound-sa-token\") pod \"4af1418f-a56e-4f26-a800-3b402b7da053\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.847850 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4af1418f-a56e-4f26-a800-3b402b7da053-ca-trust-extracted\") pod \"4af1418f-a56e-4f26-a800-3b402b7da053\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.847887 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4af1418f-a56e-4f26-a800-3b402b7da053-registry-certificates\") pod \"4af1418f-a56e-4f26-a800-3b402b7da053\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.847914 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4af1418f-a56e-4f26-a800-3b402b7da053-registry-tls\") pod \"4af1418f-a56e-4f26-a800-3b402b7da053\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.847944 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4af1418f-a56e-4f26-a800-3b402b7da053-trusted-ca\") pod \"4af1418f-a56e-4f26-a800-3b402b7da053\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.848025 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gl9h\" (UniqueName: \"kubernetes.io/projected/4af1418f-a56e-4f26-a800-3b402b7da053-kube-api-access-9gl9h\") pod \"4af1418f-a56e-4f26-a800-3b402b7da053\" (UID: \"4af1418f-a56e-4f26-a800-3b402b7da053\") " Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.848735 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4af1418f-a56e-4f26-a800-3b402b7da053-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "4af1418f-a56e-4f26-a800-3b402b7da053" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.848766 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4af1418f-a56e-4f26-a800-3b402b7da053-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "4af1418f-a56e-4f26-a800-3b402b7da053" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.855022 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4af1418f-a56e-4f26-a800-3b402b7da053-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "4af1418f-a56e-4f26-a800-3b402b7da053" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.856334 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4af1418f-a56e-4f26-a800-3b402b7da053-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "4af1418f-a56e-4f26-a800-3b402b7da053" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.856479 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4af1418f-a56e-4f26-a800-3b402b7da053-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "4af1418f-a56e-4f26-a800-3b402b7da053" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.856913 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4af1418f-a56e-4f26-a800-3b402b7da053-kube-api-access-9gl9h" (OuterVolumeSpecName: "kube-api-access-9gl9h") pod "4af1418f-a56e-4f26-a800-3b402b7da053" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053"). InnerVolumeSpecName "kube-api-access-9gl9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.860347 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "4af1418f-a56e-4f26-a800-3b402b7da053" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.864230 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4af1418f-a56e-4f26-a800-3b402b7da053-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "4af1418f-a56e-4f26-a800-3b402b7da053" (UID: "4af1418f-a56e-4f26-a800-3b402b7da053"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.949922 4809 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4af1418f-a56e-4f26-a800-3b402b7da053-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.949957 4809 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4af1418f-a56e-4f26-a800-3b402b7da053-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.950143 4809 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4af1418f-a56e-4f26-a800-3b402b7da053-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.950309 4809 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4af1418f-a56e-4f26-a800-3b402b7da053-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.950384 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4af1418f-a56e-4f26-a800-3b402b7da053-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.950462 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gl9h\" (UniqueName: \"kubernetes.io/projected/4af1418f-a56e-4f26-a800-3b402b7da053-kube-api-access-9gl9h\") on node \"crc\" DevicePath \"\"" Nov 24 07:03:15 crc kubenswrapper[4809]: I1124 07:03:15.950539 4809 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4af1418f-a56e-4f26-a800-3b402b7da053-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 07:03:16 crc kubenswrapper[4809]: I1124 07:03:16.294331 4809 generic.go:334] "Generic (PLEG): container finished" podID="4af1418f-a56e-4f26-a800-3b402b7da053" containerID="956426f3fe1032f711950ed48c6b6cff0626cabfb6e09ab8d7118943ffdfcd9e" exitCode=0 Nov 24 07:03:16 crc kubenswrapper[4809]: I1124 07:03:16.294383 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" Nov 24 07:03:16 crc kubenswrapper[4809]: I1124 07:03:16.294380 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" event={"ID":"4af1418f-a56e-4f26-a800-3b402b7da053","Type":"ContainerDied","Data":"956426f3fe1032f711950ed48c6b6cff0626cabfb6e09ab8d7118943ffdfcd9e"} Nov 24 07:03:16 crc kubenswrapper[4809]: I1124 07:03:16.294506 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-59cgv" event={"ID":"4af1418f-a56e-4f26-a800-3b402b7da053","Type":"ContainerDied","Data":"5921c15b8acc0634b00fa4668ce6b4dee98621c0f70d803226c323753698c5ce"} Nov 24 07:03:16 crc kubenswrapper[4809]: I1124 07:03:16.294536 4809 scope.go:117] "RemoveContainer" containerID="956426f3fe1032f711950ed48c6b6cff0626cabfb6e09ab8d7118943ffdfcd9e" Nov 24 07:03:16 crc kubenswrapper[4809]: I1124 07:03:16.320415 4809 scope.go:117] "RemoveContainer" containerID="956426f3fe1032f711950ed48c6b6cff0626cabfb6e09ab8d7118943ffdfcd9e" Nov 24 07:03:16 crc kubenswrapper[4809]: I1124 07:03:16.321547 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-59cgv"] Nov 24 07:03:16 crc kubenswrapper[4809]: E1124 07:03:16.322566 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"956426f3fe1032f711950ed48c6b6cff0626cabfb6e09ab8d7118943ffdfcd9e\": container with ID starting with 956426f3fe1032f711950ed48c6b6cff0626cabfb6e09ab8d7118943ffdfcd9e not found: ID does not exist" containerID="956426f3fe1032f711950ed48c6b6cff0626cabfb6e09ab8d7118943ffdfcd9e" Nov 24 07:03:16 crc kubenswrapper[4809]: I1124 07:03:16.322715 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"956426f3fe1032f711950ed48c6b6cff0626cabfb6e09ab8d7118943ffdfcd9e"} err="failed to get container status \"956426f3fe1032f711950ed48c6b6cff0626cabfb6e09ab8d7118943ffdfcd9e\": rpc error: code = NotFound desc = could not find container \"956426f3fe1032f711950ed48c6b6cff0626cabfb6e09ab8d7118943ffdfcd9e\": container with ID starting with 956426f3fe1032f711950ed48c6b6cff0626cabfb6e09ab8d7118943ffdfcd9e not found: ID does not exist" Nov 24 07:03:16 crc kubenswrapper[4809]: I1124 07:03:16.326486 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-59cgv"] Nov 24 07:03:16 crc kubenswrapper[4809]: I1124 07:03:16.898245 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4af1418f-a56e-4f26-a800-3b402b7da053" path="/var/lib/kubelet/pods/4af1418f-a56e-4f26-a800-3b402b7da053/volumes" Nov 24 07:03:48 crc kubenswrapper[4809]: I1124 07:03:48.044176 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:03:48 crc kubenswrapper[4809]: I1124 07:03:48.044852 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:04:18 crc kubenswrapper[4809]: I1124 07:04:18.042960 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:04:18 crc kubenswrapper[4809]: I1124 07:04:18.043544 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:04:48 crc kubenswrapper[4809]: I1124 07:04:48.043568 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:04:48 crc kubenswrapper[4809]: I1124 07:04:48.044457 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:04:48 crc kubenswrapper[4809]: I1124 07:04:48.044527 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 07:04:48 crc kubenswrapper[4809]: I1124 07:04:48.045603 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"24354d43179411329fa3f10b3bcaad9ac5fbb3505999caa72b922a82781a0f78"} pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:04:48 crc kubenswrapper[4809]: I1124 07:04:48.045739 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" containerID="cri-o://24354d43179411329fa3f10b3bcaad9ac5fbb3505999caa72b922a82781a0f78" gracePeriod=600 Nov 24 07:04:48 crc kubenswrapper[4809]: I1124 07:04:48.907572 4809 generic.go:334] "Generic (PLEG): container finished" podID="be8050d5-1aea-453e-a06e-87702b0e856b" containerID="24354d43179411329fa3f10b3bcaad9ac5fbb3505999caa72b922a82781a0f78" exitCode=0 Nov 24 07:04:48 crc kubenswrapper[4809]: I1124 07:04:48.907671 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerDied","Data":"24354d43179411329fa3f10b3bcaad9ac5fbb3505999caa72b922a82781a0f78"} Nov 24 07:04:48 crc kubenswrapper[4809]: I1124 07:04:48.908068 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerStarted","Data":"672cd2d0a97571eba82aabde6a902d139a7242488c7021eea3d01afc3b311c16"} Nov 24 07:04:48 crc kubenswrapper[4809]: I1124 07:04:48.908107 4809 scope.go:117] "RemoveContainer" containerID="33316ec1d7739e9081fef97a1f1a3a1c2f00b3bbab2a32058a1c6e7ff1f92aad" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.129237 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-xhsz2"] Nov 24 07:04:50 crc kubenswrapper[4809]: E1124 07:04:50.129709 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4af1418f-a56e-4f26-a800-3b402b7da053" containerName="registry" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.129721 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4af1418f-a56e-4f26-a800-3b402b7da053" containerName="registry" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.129825 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4af1418f-a56e-4f26-a800-3b402b7da053" containerName="registry" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.130202 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-xhsz2" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.133080 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.133319 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.133342 4809 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-b8m82" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.137208 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-hm272"] Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.137879 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-hm272" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.139149 4809 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-7xkr8" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.146431 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-xhsz2"] Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.151174 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-hm272"] Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.161695 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-wslw7"] Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.162311 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-wslw7" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.165102 4809 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-sbnhj" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.170344 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-wslw7"] Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.206874 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmrbl\" (UniqueName: \"kubernetes.io/projected/9de96055-2c3b-4012-a260-7b55cbb0df1b-kube-api-access-rmrbl\") pod \"cert-manager-cainjector-7f985d654d-xhsz2\" (UID: \"9de96055-2c3b-4012-a260-7b55cbb0df1b\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-xhsz2" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.206936 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsxkm\" (UniqueName: \"kubernetes.io/projected/4e03214c-a15c-4a2f-a802-ce80e9fde1c8-kube-api-access-rsxkm\") pod \"cert-manager-webhook-5655c58dd6-wslw7\" (UID: \"4e03214c-a15c-4a2f-a802-ce80e9fde1c8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-wslw7" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.207014 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqjzz\" (UniqueName: \"kubernetes.io/projected/1b39cb48-19bf-4ff6-a9e7-a2270b28ab87-kube-api-access-wqjzz\") pod \"cert-manager-5b446d88c5-hm272\" (UID: \"1b39cb48-19bf-4ff6-a9e7-a2270b28ab87\") " pod="cert-manager/cert-manager-5b446d88c5-hm272" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.309003 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqjzz\" (UniqueName: \"kubernetes.io/projected/1b39cb48-19bf-4ff6-a9e7-a2270b28ab87-kube-api-access-wqjzz\") pod \"cert-manager-5b446d88c5-hm272\" (UID: \"1b39cb48-19bf-4ff6-a9e7-a2270b28ab87\") " pod="cert-manager/cert-manager-5b446d88c5-hm272" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.309053 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmrbl\" (UniqueName: \"kubernetes.io/projected/9de96055-2c3b-4012-a260-7b55cbb0df1b-kube-api-access-rmrbl\") pod \"cert-manager-cainjector-7f985d654d-xhsz2\" (UID: \"9de96055-2c3b-4012-a260-7b55cbb0df1b\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-xhsz2" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.309102 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsxkm\" (UniqueName: \"kubernetes.io/projected/4e03214c-a15c-4a2f-a802-ce80e9fde1c8-kube-api-access-rsxkm\") pod \"cert-manager-webhook-5655c58dd6-wslw7\" (UID: \"4e03214c-a15c-4a2f-a802-ce80e9fde1c8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-wslw7" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.328905 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsxkm\" (UniqueName: \"kubernetes.io/projected/4e03214c-a15c-4a2f-a802-ce80e9fde1c8-kube-api-access-rsxkm\") pod \"cert-manager-webhook-5655c58dd6-wslw7\" (UID: \"4e03214c-a15c-4a2f-a802-ce80e9fde1c8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-wslw7" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.329105 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmrbl\" (UniqueName: \"kubernetes.io/projected/9de96055-2c3b-4012-a260-7b55cbb0df1b-kube-api-access-rmrbl\") pod \"cert-manager-cainjector-7f985d654d-xhsz2\" (UID: \"9de96055-2c3b-4012-a260-7b55cbb0df1b\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-xhsz2" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.329600 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqjzz\" (UniqueName: \"kubernetes.io/projected/1b39cb48-19bf-4ff6-a9e7-a2270b28ab87-kube-api-access-wqjzz\") pod \"cert-manager-5b446d88c5-hm272\" (UID: \"1b39cb48-19bf-4ff6-a9e7-a2270b28ab87\") " pod="cert-manager/cert-manager-5b446d88c5-hm272" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.498704 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-xhsz2" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.510090 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-hm272" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.516140 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-wslw7" Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.724467 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-xhsz2"] Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.736645 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.922334 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-xhsz2" event={"ID":"9de96055-2c3b-4012-a260-7b55cbb0df1b","Type":"ContainerStarted","Data":"900929698e03aaf872d942f67cee0cdd524cc8503a555ef01cf1601b3e2d7fe2"} Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.980664 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-wslw7"] Nov 24 07:04:50 crc kubenswrapper[4809]: W1124 07:04:50.986499 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e03214c_a15c_4a2f_a802_ce80e9fde1c8.slice/crio-07a2acbb33c87f277710b8ea0ccbcb8208a969aa8694f80b048b6ed9663da277 WatchSource:0}: Error finding container 07a2acbb33c87f277710b8ea0ccbcb8208a969aa8694f80b048b6ed9663da277: Status 404 returned error can't find the container with id 07a2acbb33c87f277710b8ea0ccbcb8208a969aa8694f80b048b6ed9663da277 Nov 24 07:04:50 crc kubenswrapper[4809]: I1124 07:04:50.991380 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-hm272"] Nov 24 07:04:50 crc kubenswrapper[4809]: W1124 07:04:50.998186 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b39cb48_19bf_4ff6_a9e7_a2270b28ab87.slice/crio-7d1139c31337178bd52035a46ab4a39ce8da319eb316f6b30c9825499972b520 WatchSource:0}: Error finding container 7d1139c31337178bd52035a46ab4a39ce8da319eb316f6b30c9825499972b520: Status 404 returned error can't find the container with id 7d1139c31337178bd52035a46ab4a39ce8da319eb316f6b30c9825499972b520 Nov 24 07:04:51 crc kubenswrapper[4809]: I1124 07:04:51.931253 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-wslw7" event={"ID":"4e03214c-a15c-4a2f-a802-ce80e9fde1c8","Type":"ContainerStarted","Data":"07a2acbb33c87f277710b8ea0ccbcb8208a969aa8694f80b048b6ed9663da277"} Nov 24 07:04:51 crc kubenswrapper[4809]: I1124 07:04:51.932602 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-hm272" event={"ID":"1b39cb48-19bf-4ff6-a9e7-a2270b28ab87","Type":"ContainerStarted","Data":"7d1139c31337178bd52035a46ab4a39ce8da319eb316f6b30c9825499972b520"} Nov 24 07:04:53 crc kubenswrapper[4809]: I1124 07:04:53.946559 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-hm272" event={"ID":"1b39cb48-19bf-4ff6-a9e7-a2270b28ab87","Type":"ContainerStarted","Data":"ec7d14b137bdcfc6e61f32fed897614039f9d0c1871ae67c658f0e3c76dcc2c7"} Nov 24 07:04:53 crc kubenswrapper[4809]: I1124 07:04:53.950089 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-xhsz2" event={"ID":"9de96055-2c3b-4012-a260-7b55cbb0df1b","Type":"ContainerStarted","Data":"387bbc26695255d1c6179ee880a559fe64b4e30bb4ccb16d660db9ad7edf30b2"} Nov 24 07:04:53 crc kubenswrapper[4809]: I1124 07:04:53.951370 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-wslw7" event={"ID":"4e03214c-a15c-4a2f-a802-ce80e9fde1c8","Type":"ContainerStarted","Data":"70f20019bb55d273d864fce5f5b62d9f1ef5208e55690e0ca87a0debd9725ef6"} Nov 24 07:04:53 crc kubenswrapper[4809]: I1124 07:04:53.951534 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-wslw7" Nov 24 07:04:53 crc kubenswrapper[4809]: I1124 07:04:53.978275 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-hm272" podStartSLOduration=1.240960868 podStartE2EDuration="3.97825694s" podCreationTimestamp="2025-11-24 07:04:50 +0000 UTC" firstStartedPulling="2025-11-24 07:04:51.000141734 +0000 UTC m=+590.900733449" lastFinishedPulling="2025-11-24 07:04:53.737437816 +0000 UTC m=+593.638029521" observedRunningTime="2025-11-24 07:04:53.961886785 +0000 UTC m=+593.862478490" watchObservedRunningTime="2025-11-24 07:04:53.97825694 +0000 UTC m=+593.878848645" Nov 24 07:04:53 crc kubenswrapper[4809]: I1124 07:04:53.995112 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-wslw7" podStartSLOduration=1.240567777 podStartE2EDuration="3.995098856s" podCreationTimestamp="2025-11-24 07:04:50 +0000 UTC" firstStartedPulling="2025-11-24 07:04:50.990563825 +0000 UTC m=+590.891155530" lastFinishedPulling="2025-11-24 07:04:53.745094914 +0000 UTC m=+593.645686609" observedRunningTime="2025-11-24 07:04:53.982218048 +0000 UTC m=+593.882809753" watchObservedRunningTime="2025-11-24 07:04:53.995098856 +0000 UTC m=+593.895690561" Nov 24 07:04:53 crc kubenswrapper[4809]: I1124 07:04:53.996016 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-xhsz2" podStartSLOduration=2.039536504 podStartE2EDuration="3.996010272s" podCreationTimestamp="2025-11-24 07:04:50 +0000 UTC" firstStartedPulling="2025-11-24 07:04:50.736385579 +0000 UTC m=+590.636977284" lastFinishedPulling="2025-11-24 07:04:52.692859347 +0000 UTC m=+592.593451052" observedRunningTime="2025-11-24 07:04:53.99334657 +0000 UTC m=+593.893938275" watchObservedRunningTime="2025-11-24 07:04:53.996010272 +0000 UTC m=+593.896601977" Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.521150 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-wslw7" Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.731799 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-kpqhl"] Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.732430 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovn-controller" containerID="cri-o://54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed" gracePeriod=30 Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.732517 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="nbdb" containerID="cri-o://62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7" gracePeriod=30 Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.732623 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="northd" containerID="cri-o://787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a" gracePeriod=30 Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.732698 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844" gracePeriod=30 Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.732767 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="kube-rbac-proxy-node" containerID="cri-o://0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844" gracePeriod=30 Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.732824 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovn-acl-logging" containerID="cri-o://88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249" gracePeriod=30 Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.732952 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="sbdb" containerID="cri-o://a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c" gracePeriod=30 Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.818719 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovnkube-controller" containerID="cri-o://aac5161cc51c14f1071e02e84f4cd9e1e7656f9a466ea5088b41c34d2687f929" gracePeriod=30 Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.996081 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kpqhl_74622ad8-44b0-44a7-afbd-e19bb9635520/ovnkube-controller/3.log" Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.998156 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kpqhl_74622ad8-44b0-44a7-afbd-e19bb9635520/ovn-acl-logging/0.log" Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.998601 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kpqhl_74622ad8-44b0-44a7-afbd-e19bb9635520/ovn-controller/0.log" Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.998930 4809 generic.go:334] "Generic (PLEG): container finished" podID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerID="aac5161cc51c14f1071e02e84f4cd9e1e7656f9a466ea5088b41c34d2687f929" exitCode=0 Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.998977 4809 generic.go:334] "Generic (PLEG): container finished" podID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerID="cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844" exitCode=0 Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.998987 4809 generic.go:334] "Generic (PLEG): container finished" podID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerID="0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844" exitCode=0 Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.998997 4809 generic.go:334] "Generic (PLEG): container finished" podID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerID="88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249" exitCode=143 Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.999006 4809 generic.go:334] "Generic (PLEG): container finished" podID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerID="54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed" exitCode=143 Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.999063 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerDied","Data":"aac5161cc51c14f1071e02e84f4cd9e1e7656f9a466ea5088b41c34d2687f929"} Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.999094 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerDied","Data":"cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844"} Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.999107 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerDied","Data":"0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844"} Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.999120 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerDied","Data":"88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249"} Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.999132 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerDied","Data":"54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed"} Nov 24 07:05:00 crc kubenswrapper[4809]: I1124 07:05:00.999152 4809 scope.go:117] "RemoveContainer" containerID="7f3517cb223a94e59385d72b6844b837a34cb285128c89db3ef6f259a4d505f3" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.002590 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5xvl8_3c932894-56b4-4303-8176-d12b12686b09/kube-multus/2.log" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.003025 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5xvl8_3c932894-56b4-4303-8176-d12b12686b09/kube-multus/1.log" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.003189 4809 generic.go:334] "Generic (PLEG): container finished" podID="3c932894-56b4-4303-8176-d12b12686b09" containerID="4f0050e9681c436b5268f8f5d15039d7e02d064e44ac5fecf5a52b0ad3cf3a81" exitCode=2 Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.003258 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5xvl8" event={"ID":"3c932894-56b4-4303-8176-d12b12686b09","Type":"ContainerDied","Data":"4f0050e9681c436b5268f8f5d15039d7e02d064e44ac5fecf5a52b0ad3cf3a81"} Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.003722 4809 scope.go:117] "RemoveContainer" containerID="4f0050e9681c436b5268f8f5d15039d7e02d064e44ac5fecf5a52b0ad3cf3a81" Nov 24 07:05:01 crc kubenswrapper[4809]: E1124 07:05:01.003887 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-5xvl8_openshift-multus(3c932894-56b4-4303-8176-d12b12686b09)\"" pod="openshift-multus/multus-5xvl8" podUID="3c932894-56b4-4303-8176-d12b12686b09" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.040059 4809 scope.go:117] "RemoveContainer" containerID="7f3517cb223a94e59385d72b6844b837a34cb285128c89db3ef6f259a4d505f3" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.098425 4809 scope.go:117] "RemoveContainer" containerID="d68cb2e2c7982f3c55f7d9263588e648ba155a930efc0266051aaa628ccb9db7" Nov 24 07:05:01 crc kubenswrapper[4809]: E1124 07:05:01.098626 4809 log.go:32] "RemoveContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to delete container k8s_ovnkube-controller_ovnkube-node-kpqhl_openshift-ovn-kubernetes_74622ad8-44b0-44a7-afbd-e19bb9635520_3 in pod sandbox be0abc73293942a2e9751f120dda9b69eb20a33702af201e3c4e611d18e7705f from index: no such id: '7f3517cb223a94e59385d72b6844b837a34cb285128c89db3ef6f259a4d505f3'" containerID="7f3517cb223a94e59385d72b6844b837a34cb285128c89db3ef6f259a4d505f3" Nov 24 07:05:01 crc kubenswrapper[4809]: E1124 07:05:01.098693 4809 kuberuntime_gc.go:150] "Failed to remove container" err="rpc error: code = Unknown desc = failed to delete container k8s_ovnkube-controller_ovnkube-node-kpqhl_openshift-ovn-kubernetes_74622ad8-44b0-44a7-afbd-e19bb9635520_3 in pod sandbox be0abc73293942a2e9751f120dda9b69eb20a33702af201e3c4e611d18e7705f from index: no such id: '7f3517cb223a94e59385d72b6844b837a34cb285128c89db3ef6f259a4d505f3'" containerID="7f3517cb223a94e59385d72b6844b837a34cb285128c89db3ef6f259a4d505f3" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.098722 4809 scope.go:117] "RemoveContainer" containerID="d68cb2e2c7982f3c55f7d9263588e648ba155a930efc0266051aaa628ccb9db7" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.114910 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kpqhl_74622ad8-44b0-44a7-afbd-e19bb9635520/ovn-acl-logging/0.log" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.115343 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kpqhl_74622ad8-44b0-44a7-afbd-e19bb9635520/ovn-controller/0.log" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.115828 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 07:05:01 crc kubenswrapper[4809]: E1124 07:05:01.127660 4809 log.go:32] "RemoveContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to delete container k8s_kube-multus_multus-5xvl8_openshift-multus_3c932894-56b4-4303-8176-d12b12686b09_1 in pod sandbox 983884fed6e5b709a8aa38ebb319995e7a01ab7079472e5e4ced48036befcad7 from index: no such id: 'd68cb2e2c7982f3c55f7d9263588e648ba155a930efc0266051aaa628ccb9db7'" containerID="d68cb2e2c7982f3c55f7d9263588e648ba155a930efc0266051aaa628ccb9db7" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.127697 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d68cb2e2c7982f3c55f7d9263588e648ba155a930efc0266051aaa628ccb9db7"} err="rpc error: code = Unknown desc = failed to delete container k8s_kube-multus_multus-5xvl8_openshift-multus_3c932894-56b4-4303-8176-d12b12686b09_1 in pod sandbox 983884fed6e5b709a8aa38ebb319995e7a01ab7079472e5e4ced48036befcad7 from index: no such id: 'd68cb2e2c7982f3c55f7d9263588e648ba155a930efc0266051aaa628ccb9db7'" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.165211 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-run-ovn-kubernetes\") pod \"74622ad8-44b0-44a7-afbd-e19bb9635520\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.165284 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-run-openvswitch\") pod \"74622ad8-44b0-44a7-afbd-e19bb9635520\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.165343 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-etc-openvswitch\") pod \"74622ad8-44b0-44a7-afbd-e19bb9635520\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.165403 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/74622ad8-44b0-44a7-afbd-e19bb9635520-ovn-node-metrics-cert\") pod \"74622ad8-44b0-44a7-afbd-e19bb9635520\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.165426 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-run-netns\") pod \"74622ad8-44b0-44a7-afbd-e19bb9635520\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.165445 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-systemd-units\") pod \"74622ad8-44b0-44a7-afbd-e19bb9635520\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.165503 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7hjg\" (UniqueName: \"kubernetes.io/projected/74622ad8-44b0-44a7-afbd-e19bb9635520-kube-api-access-t7hjg\") pod \"74622ad8-44b0-44a7-afbd-e19bb9635520\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.165530 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/74622ad8-44b0-44a7-afbd-e19bb9635520-env-overrides\") pod \"74622ad8-44b0-44a7-afbd-e19bb9635520\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.165580 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-slash\") pod \"74622ad8-44b0-44a7-afbd-e19bb9635520\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.165607 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-cni-netd\") pod \"74622ad8-44b0-44a7-afbd-e19bb9635520\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.165668 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-var-lib-openvswitch\") pod \"74622ad8-44b0-44a7-afbd-e19bb9635520\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.165692 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-run-ovn\") pod \"74622ad8-44b0-44a7-afbd-e19bb9635520\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.165746 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-var-lib-cni-networks-ovn-kubernetes\") pod \"74622ad8-44b0-44a7-afbd-e19bb9635520\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.165771 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-cni-bin\") pod \"74622ad8-44b0-44a7-afbd-e19bb9635520\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.165823 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/74622ad8-44b0-44a7-afbd-e19bb9635520-ovnkube-config\") pod \"74622ad8-44b0-44a7-afbd-e19bb9635520\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.165847 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-log-socket\") pod \"74622ad8-44b0-44a7-afbd-e19bb9635520\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.165914 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-kubelet\") pod \"74622ad8-44b0-44a7-afbd-e19bb9635520\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.165945 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/74622ad8-44b0-44a7-afbd-e19bb9635520-ovnkube-script-lib\") pod \"74622ad8-44b0-44a7-afbd-e19bb9635520\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.166008 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-run-systemd\") pod \"74622ad8-44b0-44a7-afbd-e19bb9635520\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.166031 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-node-log\") pod \"74622ad8-44b0-44a7-afbd-e19bb9635520\" (UID: \"74622ad8-44b0-44a7-afbd-e19bb9635520\") " Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.166468 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-node-log" (OuterVolumeSpecName: "node-log") pod "74622ad8-44b0-44a7-afbd-e19bb9635520" (UID: "74622ad8-44b0-44a7-afbd-e19bb9635520"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.166555 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "74622ad8-44b0-44a7-afbd-e19bb9635520" (UID: "74622ad8-44b0-44a7-afbd-e19bb9635520"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.166619 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "74622ad8-44b0-44a7-afbd-e19bb9635520" (UID: "74622ad8-44b0-44a7-afbd-e19bb9635520"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.166644 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "74622ad8-44b0-44a7-afbd-e19bb9635520" (UID: "74622ad8-44b0-44a7-afbd-e19bb9635520"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167009 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zrmrp"] Nov 24 07:05:01 crc kubenswrapper[4809]: E1124 07:05:01.167223 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovn-acl-logging" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167242 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovn-acl-logging" Nov 24 07:05:01 crc kubenswrapper[4809]: E1124 07:05:01.167252 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="nbdb" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167259 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="nbdb" Nov 24 07:05:01 crc kubenswrapper[4809]: E1124 07:05:01.167269 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovnkube-controller" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167275 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovnkube-controller" Nov 24 07:05:01 crc kubenswrapper[4809]: E1124 07:05:01.167283 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="kubecfg-setup" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167288 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="kubecfg-setup" Nov 24 07:05:01 crc kubenswrapper[4809]: E1124 07:05:01.167296 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="kube-rbac-proxy-node" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167302 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="kube-rbac-proxy-node" Nov 24 07:05:01 crc kubenswrapper[4809]: E1124 07:05:01.167314 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovnkube-controller" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167319 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovnkube-controller" Nov 24 07:05:01 crc kubenswrapper[4809]: E1124 07:05:01.167326 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovn-controller" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167332 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovn-controller" Nov 24 07:05:01 crc kubenswrapper[4809]: E1124 07:05:01.167341 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="sbdb" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167347 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="sbdb" Nov 24 07:05:01 crc kubenswrapper[4809]: E1124 07:05:01.167355 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovnkube-controller" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167361 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovnkube-controller" Nov 24 07:05:01 crc kubenswrapper[4809]: E1124 07:05:01.167368 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="northd" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167373 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="northd" Nov 24 07:05:01 crc kubenswrapper[4809]: E1124 07:05:01.167382 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovnkube-controller" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167389 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovnkube-controller" Nov 24 07:05:01 crc kubenswrapper[4809]: E1124 07:05:01.167397 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167402 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167484 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovnkube-controller" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167493 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovnkube-controller" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167501 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="northd" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167510 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovnkube-controller" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167516 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="sbdb" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167523 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167531 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovnkube-controller" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167538 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="nbdb" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167545 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovn-controller" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167552 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="kube-rbac-proxy-node" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167560 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovnkube-controller" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167567 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovn-acl-logging" Nov 24 07:05:01 crc kubenswrapper[4809]: E1124 07:05:01.167658 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovnkube-controller" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.167665 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerName="ovnkube-controller" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.168131 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "74622ad8-44b0-44a7-afbd-e19bb9635520" (UID: "74622ad8-44b0-44a7-afbd-e19bb9635520"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.168176 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-slash" (OuterVolumeSpecName: "host-slash") pod "74622ad8-44b0-44a7-afbd-e19bb9635520" (UID: "74622ad8-44b0-44a7-afbd-e19bb9635520"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.168188 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-log-socket" (OuterVolumeSpecName: "log-socket") pod "74622ad8-44b0-44a7-afbd-e19bb9635520" (UID: "74622ad8-44b0-44a7-afbd-e19bb9635520"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.168215 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "74622ad8-44b0-44a7-afbd-e19bb9635520" (UID: "74622ad8-44b0-44a7-afbd-e19bb9635520"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.168245 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "74622ad8-44b0-44a7-afbd-e19bb9635520" (UID: "74622ad8-44b0-44a7-afbd-e19bb9635520"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.168271 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "74622ad8-44b0-44a7-afbd-e19bb9635520" (UID: "74622ad8-44b0-44a7-afbd-e19bb9635520"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.168297 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "74622ad8-44b0-44a7-afbd-e19bb9635520" (UID: "74622ad8-44b0-44a7-afbd-e19bb9635520"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.168287 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "74622ad8-44b0-44a7-afbd-e19bb9635520" (UID: "74622ad8-44b0-44a7-afbd-e19bb9635520"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.168282 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "74622ad8-44b0-44a7-afbd-e19bb9635520" (UID: "74622ad8-44b0-44a7-afbd-e19bb9635520"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.168301 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "74622ad8-44b0-44a7-afbd-e19bb9635520" (UID: "74622ad8-44b0-44a7-afbd-e19bb9635520"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.168661 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74622ad8-44b0-44a7-afbd-e19bb9635520-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "74622ad8-44b0-44a7-afbd-e19bb9635520" (UID: "74622ad8-44b0-44a7-afbd-e19bb9635520"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.168823 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74622ad8-44b0-44a7-afbd-e19bb9635520-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "74622ad8-44b0-44a7-afbd-e19bb9635520" (UID: "74622ad8-44b0-44a7-afbd-e19bb9635520"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.169315 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.169536 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74622ad8-44b0-44a7-afbd-e19bb9635520-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "74622ad8-44b0-44a7-afbd-e19bb9635520" (UID: "74622ad8-44b0-44a7-afbd-e19bb9635520"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.174011 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74622ad8-44b0-44a7-afbd-e19bb9635520-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "74622ad8-44b0-44a7-afbd-e19bb9635520" (UID: "74622ad8-44b0-44a7-afbd-e19bb9635520"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.174791 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74622ad8-44b0-44a7-afbd-e19bb9635520-kube-api-access-t7hjg" (OuterVolumeSpecName: "kube-api-access-t7hjg") pod "74622ad8-44b0-44a7-afbd-e19bb9635520" (UID: "74622ad8-44b0-44a7-afbd-e19bb9635520"). InnerVolumeSpecName "kube-api-access-t7hjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.188680 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "74622ad8-44b0-44a7-afbd-e19bb9635520" (UID: "74622ad8-44b0-44a7-afbd-e19bb9635520"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.267919 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-host-kubelet\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.268017 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-host-cni-bin\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.268099 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-etc-openvswitch\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.268184 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.268264 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-run-openvswitch\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.268339 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-run-systemd\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.268385 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-ovnkube-config\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.268417 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-systemd-units\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.268478 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-env-overrides\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.268515 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfjw5\" (UniqueName: \"kubernetes.io/projected/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-kube-api-access-hfjw5\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.268565 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-ovnkube-script-lib\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.268615 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-log-socket\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.268657 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-ovn-node-metrics-cert\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.268701 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-host-slash\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.268733 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-host-run-netns\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.268834 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-node-log\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.268901 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-run-ovn\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.268938 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-host-run-ovn-kubernetes\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.269066 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-var-lib-openvswitch\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.269100 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-host-cni-netd\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.269217 4809 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.269241 4809 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.269257 4809 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.269268 4809 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.269277 4809 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.269287 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/74622ad8-44b0-44a7-afbd-e19bb9635520-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.269297 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7hjg\" (UniqueName: \"kubernetes.io/projected/74622ad8-44b0-44a7-afbd-e19bb9635520-kube-api-access-t7hjg\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.269308 4809 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/74622ad8-44b0-44a7-afbd-e19bb9635520-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.269320 4809 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-slash\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.269329 4809 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.269339 4809 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.269395 4809 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.269407 4809 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.269442 4809 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.269453 4809 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/74622ad8-44b0-44a7-afbd-e19bb9635520-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.269464 4809 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-log-socket\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.269476 4809 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.269489 4809 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/74622ad8-44b0-44a7-afbd-e19bb9635520-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.269502 4809 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.269513 4809 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/74622ad8-44b0-44a7-afbd-e19bb9635520-node-log\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.370835 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-ovn-node-metrics-cert\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.370942 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-host-slash\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.371036 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-host-run-netns\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.371086 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-node-log\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.371132 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-run-ovn\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.371137 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-host-slash\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.371261 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-node-log\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.371288 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-host-run-netns\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.371400 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-run-ovn\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.371288 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-host-run-ovn-kubernetes\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.371528 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-host-run-ovn-kubernetes\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.371654 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-var-lib-openvswitch\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.371701 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-host-cni-netd\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.371746 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-host-kubelet\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.371775 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-host-cni-bin\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.371805 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-etc-openvswitch\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.371844 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-host-cni-bin\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.371862 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-etc-openvswitch\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.371805 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-var-lib-openvswitch\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.371877 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-host-cni-netd\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.371895 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.371844 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.372017 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-run-openvswitch\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.372099 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-run-systemd\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.372163 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-ovnkube-config\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.372204 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-systemd-units\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.372253 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-env-overrides\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.372305 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfjw5\" (UniqueName: \"kubernetes.io/projected/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-kube-api-access-hfjw5\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.372364 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-ovnkube-script-lib\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.372415 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-log-socket\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.372528 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-log-socket\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.372590 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-run-openvswitch\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.372645 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-run-systemd\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.373709 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-systemd-units\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.374091 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-ovnkube-config\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.374286 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-host-kubelet\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.374904 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-env-overrides\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.375030 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-ovn-node-metrics-cert\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.375798 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-ovnkube-script-lib\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.402476 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfjw5\" (UniqueName: \"kubernetes.io/projected/61a1ad5a-317f-481d-ab9f-8d2e4cee860d-kube-api-access-hfjw5\") pod \"ovnkube-node-zrmrp\" (UID: \"61a1ad5a-317f-481d-ab9f-8d2e4cee860d\") " pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: I1124 07:05:01.490152 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:01 crc kubenswrapper[4809]: W1124 07:05:01.518048 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61a1ad5a_317f_481d_ab9f_8d2e4cee860d.slice/crio-0da2d3311a6e1e60a8c1f9aded1bff13449df8f02616bbaf5bef4e1184e5392f WatchSource:0}: Error finding container 0da2d3311a6e1e60a8c1f9aded1bff13449df8f02616bbaf5bef4e1184e5392f: Status 404 returned error can't find the container with id 0da2d3311a6e1e60a8c1f9aded1bff13449df8f02616bbaf5bef4e1184e5392f Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.011204 4809 generic.go:334] "Generic (PLEG): container finished" podID="61a1ad5a-317f-481d-ab9f-8d2e4cee860d" containerID="f54039e5d5065f4a1834fbf8cf4244f0b1684a8369c1aa9847f0f09be43efa61" exitCode=0 Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.011336 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" event={"ID":"61a1ad5a-317f-481d-ab9f-8d2e4cee860d","Type":"ContainerDied","Data":"f54039e5d5065f4a1834fbf8cf4244f0b1684a8369c1aa9847f0f09be43efa61"} Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.011607 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" event={"ID":"61a1ad5a-317f-481d-ab9f-8d2e4cee860d","Type":"ContainerStarted","Data":"0da2d3311a6e1e60a8c1f9aded1bff13449df8f02616bbaf5bef4e1184e5392f"} Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.020215 4809 generic.go:334] "Generic (PLEG): container finished" podID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerID="a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c" exitCode=0 Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.020248 4809 generic.go:334] "Generic (PLEG): container finished" podID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerID="62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7" exitCode=0 Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.020261 4809 generic.go:334] "Generic (PLEG): container finished" podID="74622ad8-44b0-44a7-afbd-e19bb9635520" containerID="787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a" exitCode=0 Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.020304 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.020300 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerDied","Data":"a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c"} Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.020348 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerDied","Data":"62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7"} Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.020363 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerDied","Data":"787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a"} Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.020377 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kpqhl" event={"ID":"74622ad8-44b0-44a7-afbd-e19bb9635520","Type":"ContainerDied","Data":"be0abc73293942a2e9751f120dda9b69eb20a33702af201e3c4e611d18e7705f"} Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.020395 4809 scope.go:117] "RemoveContainer" containerID="aac5161cc51c14f1071e02e84f4cd9e1e7656f9a466ea5088b41c34d2687f929" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.023801 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5xvl8_3c932894-56b4-4303-8176-d12b12686b09/kube-multus/2.log" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.042172 4809 scope.go:117] "RemoveContainer" containerID="a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.092294 4809 scope.go:117] "RemoveContainer" containerID="62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.115442 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-kpqhl"] Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.126069 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-kpqhl"] Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.135375 4809 scope.go:117] "RemoveContainer" containerID="787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.157480 4809 scope.go:117] "RemoveContainer" containerID="cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.181938 4809 scope.go:117] "RemoveContainer" containerID="0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.201123 4809 scope.go:117] "RemoveContainer" containerID="88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.227298 4809 scope.go:117] "RemoveContainer" containerID="54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.245794 4809 scope.go:117] "RemoveContainer" containerID="9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.276125 4809 scope.go:117] "RemoveContainer" containerID="aac5161cc51c14f1071e02e84f4cd9e1e7656f9a466ea5088b41c34d2687f929" Nov 24 07:05:02 crc kubenswrapper[4809]: E1124 07:05:02.276487 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aac5161cc51c14f1071e02e84f4cd9e1e7656f9a466ea5088b41c34d2687f929\": container with ID starting with aac5161cc51c14f1071e02e84f4cd9e1e7656f9a466ea5088b41c34d2687f929 not found: ID does not exist" containerID="aac5161cc51c14f1071e02e84f4cd9e1e7656f9a466ea5088b41c34d2687f929" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.276519 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aac5161cc51c14f1071e02e84f4cd9e1e7656f9a466ea5088b41c34d2687f929"} err="failed to get container status \"aac5161cc51c14f1071e02e84f4cd9e1e7656f9a466ea5088b41c34d2687f929\": rpc error: code = NotFound desc = could not find container \"aac5161cc51c14f1071e02e84f4cd9e1e7656f9a466ea5088b41c34d2687f929\": container with ID starting with aac5161cc51c14f1071e02e84f4cd9e1e7656f9a466ea5088b41c34d2687f929 not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.276539 4809 scope.go:117] "RemoveContainer" containerID="a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c" Nov 24 07:05:02 crc kubenswrapper[4809]: E1124 07:05:02.276957 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\": container with ID starting with a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c not found: ID does not exist" containerID="a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.277009 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c"} err="failed to get container status \"a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\": rpc error: code = NotFound desc = could not find container \"a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\": container with ID starting with a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.277037 4809 scope.go:117] "RemoveContainer" containerID="62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7" Nov 24 07:05:02 crc kubenswrapper[4809]: E1124 07:05:02.277388 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\": container with ID starting with 62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7 not found: ID does not exist" containerID="62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.277412 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7"} err="failed to get container status \"62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\": rpc error: code = NotFound desc = could not find container \"62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\": container with ID starting with 62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7 not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.277427 4809 scope.go:117] "RemoveContainer" containerID="787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a" Nov 24 07:05:02 crc kubenswrapper[4809]: E1124 07:05:02.277684 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\": container with ID starting with 787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a not found: ID does not exist" containerID="787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.277712 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a"} err="failed to get container status \"787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\": rpc error: code = NotFound desc = could not find container \"787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\": container with ID starting with 787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.277728 4809 scope.go:117] "RemoveContainer" containerID="cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844" Nov 24 07:05:02 crc kubenswrapper[4809]: E1124 07:05:02.278125 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\": container with ID starting with cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844 not found: ID does not exist" containerID="cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.278150 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844"} err="failed to get container status \"cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\": rpc error: code = NotFound desc = could not find container \"cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\": container with ID starting with cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844 not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.278162 4809 scope.go:117] "RemoveContainer" containerID="0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844" Nov 24 07:05:02 crc kubenswrapper[4809]: E1124 07:05:02.278477 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\": container with ID starting with 0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844 not found: ID does not exist" containerID="0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.278505 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844"} err="failed to get container status \"0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\": rpc error: code = NotFound desc = could not find container \"0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\": container with ID starting with 0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844 not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.278520 4809 scope.go:117] "RemoveContainer" containerID="88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249" Nov 24 07:05:02 crc kubenswrapper[4809]: E1124 07:05:02.278753 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\": container with ID starting with 88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249 not found: ID does not exist" containerID="88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.278776 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249"} err="failed to get container status \"88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\": rpc error: code = NotFound desc = could not find container \"88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\": container with ID starting with 88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249 not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.278788 4809 scope.go:117] "RemoveContainer" containerID="54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed" Nov 24 07:05:02 crc kubenswrapper[4809]: E1124 07:05:02.279227 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\": container with ID starting with 54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed not found: ID does not exist" containerID="54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.279247 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed"} err="failed to get container status \"54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\": rpc error: code = NotFound desc = could not find container \"54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\": container with ID starting with 54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.279265 4809 scope.go:117] "RemoveContainer" containerID="9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855" Nov 24 07:05:02 crc kubenswrapper[4809]: E1124 07:05:02.279502 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\": container with ID starting with 9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855 not found: ID does not exist" containerID="9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.279530 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855"} err="failed to get container status \"9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\": rpc error: code = NotFound desc = could not find container \"9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\": container with ID starting with 9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855 not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.279550 4809 scope.go:117] "RemoveContainer" containerID="aac5161cc51c14f1071e02e84f4cd9e1e7656f9a466ea5088b41c34d2687f929" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.279857 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aac5161cc51c14f1071e02e84f4cd9e1e7656f9a466ea5088b41c34d2687f929"} err="failed to get container status \"aac5161cc51c14f1071e02e84f4cd9e1e7656f9a466ea5088b41c34d2687f929\": rpc error: code = NotFound desc = could not find container \"aac5161cc51c14f1071e02e84f4cd9e1e7656f9a466ea5088b41c34d2687f929\": container with ID starting with aac5161cc51c14f1071e02e84f4cd9e1e7656f9a466ea5088b41c34d2687f929 not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.279880 4809 scope.go:117] "RemoveContainer" containerID="a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.280413 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c"} err="failed to get container status \"a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\": rpc error: code = NotFound desc = could not find container \"a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\": container with ID starting with a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.280436 4809 scope.go:117] "RemoveContainer" containerID="62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.280709 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7"} err="failed to get container status \"62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\": rpc error: code = NotFound desc = could not find container \"62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\": container with ID starting with 62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7 not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.280731 4809 scope.go:117] "RemoveContainer" containerID="787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.281021 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a"} err="failed to get container status \"787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\": rpc error: code = NotFound desc = could not find container \"787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\": container with ID starting with 787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.281040 4809 scope.go:117] "RemoveContainer" containerID="cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.281294 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844"} err="failed to get container status \"cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\": rpc error: code = NotFound desc = could not find container \"cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\": container with ID starting with cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844 not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.281313 4809 scope.go:117] "RemoveContainer" containerID="0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.281560 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844"} err="failed to get container status \"0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\": rpc error: code = NotFound desc = could not find container \"0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\": container with ID starting with 0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844 not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.281578 4809 scope.go:117] "RemoveContainer" containerID="88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.281887 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249"} err="failed to get container status \"88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\": rpc error: code = NotFound desc = could not find container \"88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\": container with ID starting with 88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249 not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.281907 4809 scope.go:117] "RemoveContainer" containerID="54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.282164 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed"} err="failed to get container status \"54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\": rpc error: code = NotFound desc = could not find container \"54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\": container with ID starting with 54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.282185 4809 scope.go:117] "RemoveContainer" containerID="9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.282425 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855"} err="failed to get container status \"9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\": rpc error: code = NotFound desc = could not find container \"9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\": container with ID starting with 9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855 not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.282446 4809 scope.go:117] "RemoveContainer" containerID="aac5161cc51c14f1071e02e84f4cd9e1e7656f9a466ea5088b41c34d2687f929" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.282672 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aac5161cc51c14f1071e02e84f4cd9e1e7656f9a466ea5088b41c34d2687f929"} err="failed to get container status \"aac5161cc51c14f1071e02e84f4cd9e1e7656f9a466ea5088b41c34d2687f929\": rpc error: code = NotFound desc = could not find container \"aac5161cc51c14f1071e02e84f4cd9e1e7656f9a466ea5088b41c34d2687f929\": container with ID starting with aac5161cc51c14f1071e02e84f4cd9e1e7656f9a466ea5088b41c34d2687f929 not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.282710 4809 scope.go:117] "RemoveContainer" containerID="a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.282905 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c"} err="failed to get container status \"a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\": rpc error: code = NotFound desc = could not find container \"a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c\": container with ID starting with a9527a4350f41a4f89df4912b1a05cc8a4b108a09d8339fe402667191f87fa2c not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.282923 4809 scope.go:117] "RemoveContainer" containerID="62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.283161 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7"} err="failed to get container status \"62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\": rpc error: code = NotFound desc = could not find container \"62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7\": container with ID starting with 62d4845fb8396d5b03badc9337476606d358c8bffe431e9a3372bc02a23433c7 not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.283181 4809 scope.go:117] "RemoveContainer" containerID="787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.283379 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a"} err="failed to get container status \"787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\": rpc error: code = NotFound desc = could not find container \"787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a\": container with ID starting with 787af784b5adc19418636e6a69fec8f435eb8eb285f7ffea90e60a687f8b728a not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.283396 4809 scope.go:117] "RemoveContainer" containerID="cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.283699 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844"} err="failed to get container status \"cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\": rpc error: code = NotFound desc = could not find container \"cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844\": container with ID starting with cc8dccd8ce509d9c2f902fe6d14c220479ac4107b116d6feded982a45c20c844 not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.283719 4809 scope.go:117] "RemoveContainer" containerID="0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.283921 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844"} err="failed to get container status \"0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\": rpc error: code = NotFound desc = could not find container \"0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844\": container with ID starting with 0e4519732fd3a354028aff397bf8dd1e6947f995895a77cb9ec21615b9485844 not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.283942 4809 scope.go:117] "RemoveContainer" containerID="88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.284150 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249"} err="failed to get container status \"88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\": rpc error: code = NotFound desc = could not find container \"88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249\": container with ID starting with 88812ff0ba201402279985488853fec018d4513952dc21f3399e7f2c433ff249 not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.284170 4809 scope.go:117] "RemoveContainer" containerID="54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.284378 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed"} err="failed to get container status \"54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\": rpc error: code = NotFound desc = could not find container \"54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed\": container with ID starting with 54f060d636481b45eaf0e2f07645a2d540fa86d8cfe36c7939dcf1d65658b6ed not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.284399 4809 scope.go:117] "RemoveContainer" containerID="9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.285129 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855"} err="failed to get container status \"9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\": rpc error: code = NotFound desc = could not find container \"9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855\": container with ID starting with 9feb77bf0fbca8cc5a5f96fa2b97c6c2d6eb6d9a9ff024302048f95c7e5e3855 not found: ID does not exist" Nov 24 07:05:02 crc kubenswrapper[4809]: I1124 07:05:02.898090 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74622ad8-44b0-44a7-afbd-e19bb9635520" path="/var/lib/kubelet/pods/74622ad8-44b0-44a7-afbd-e19bb9635520/volumes" Nov 24 07:05:03 crc kubenswrapper[4809]: I1124 07:05:03.037103 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" event={"ID":"61a1ad5a-317f-481d-ab9f-8d2e4cee860d","Type":"ContainerStarted","Data":"dbfd80bac0a1200eeb1104af1c8cff20cbf14e76295db1695f3b5cc316bdcb37"} Nov 24 07:05:03 crc kubenswrapper[4809]: I1124 07:05:03.038160 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" event={"ID":"61a1ad5a-317f-481d-ab9f-8d2e4cee860d","Type":"ContainerStarted","Data":"4860b77738ea569eeee660fb3c767d16c459685623ba8daf1d8a9cf6a0b37d18"} Nov 24 07:05:03 crc kubenswrapper[4809]: I1124 07:05:03.038203 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" event={"ID":"61a1ad5a-317f-481d-ab9f-8d2e4cee860d","Type":"ContainerStarted","Data":"22892571e0c793a5f420cb4103b2fa2c922d1b599697cc07cd737d302f0169b0"} Nov 24 07:05:03 crc kubenswrapper[4809]: I1124 07:05:03.038225 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" event={"ID":"61a1ad5a-317f-481d-ab9f-8d2e4cee860d","Type":"ContainerStarted","Data":"4e1509c7587da2e0d702209083a505473bda18e0b86a31c56370520b5835737e"} Nov 24 07:05:03 crc kubenswrapper[4809]: I1124 07:05:03.038243 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" event={"ID":"61a1ad5a-317f-481d-ab9f-8d2e4cee860d","Type":"ContainerStarted","Data":"c62c8e90c6db65302ac5859328a6477dfe60afb5a5db8dcbb9f0c1e2953f1556"} Nov 24 07:05:03 crc kubenswrapper[4809]: I1124 07:05:03.038261 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" event={"ID":"61a1ad5a-317f-481d-ab9f-8d2e4cee860d","Type":"ContainerStarted","Data":"4cdae1d8d9225622288f4277963f8b3ecbfe0ab014f52b39b225acdf8cf7cbbe"} Nov 24 07:05:06 crc kubenswrapper[4809]: I1124 07:05:06.061072 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" event={"ID":"61a1ad5a-317f-481d-ab9f-8d2e4cee860d","Type":"ContainerStarted","Data":"7d4e002821c0f13ebbf6e9af8475854ae0072d93e294cc93b9edec1f845ae010"} Nov 24 07:05:08 crc kubenswrapper[4809]: I1124 07:05:08.084020 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" event={"ID":"61a1ad5a-317f-481d-ab9f-8d2e4cee860d","Type":"ContainerStarted","Data":"23166312dd2e39489ce9ace7696d438fe33021c99847793f304907c7d45bff17"} Nov 24 07:05:08 crc kubenswrapper[4809]: I1124 07:05:08.084571 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:08 crc kubenswrapper[4809]: I1124 07:05:08.084588 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:08 crc kubenswrapper[4809]: I1124 07:05:08.133811 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" podStartSLOduration=7.133787116 podStartE2EDuration="7.133787116s" podCreationTimestamp="2025-11-24 07:05:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:05:08.131613828 +0000 UTC m=+608.032205553" watchObservedRunningTime="2025-11-24 07:05:08.133787116 +0000 UTC m=+608.034378871" Nov 24 07:05:08 crc kubenswrapper[4809]: I1124 07:05:08.138219 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:09 crc kubenswrapper[4809]: I1124 07:05:09.094227 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:09 crc kubenswrapper[4809]: I1124 07:05:09.138883 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:14 crc kubenswrapper[4809]: I1124 07:05:14.891885 4809 scope.go:117] "RemoveContainer" containerID="4f0050e9681c436b5268f8f5d15039d7e02d064e44ac5fecf5a52b0ad3cf3a81" Nov 24 07:05:14 crc kubenswrapper[4809]: E1124 07:05:14.893827 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-5xvl8_openshift-multus(3c932894-56b4-4303-8176-d12b12686b09)\"" pod="openshift-multus/multus-5xvl8" podUID="3c932894-56b4-4303-8176-d12b12686b09" Nov 24 07:05:29 crc kubenswrapper[4809]: I1124 07:05:29.891652 4809 scope.go:117] "RemoveContainer" containerID="4f0050e9681c436b5268f8f5d15039d7e02d064e44ac5fecf5a52b0ad3cf3a81" Nov 24 07:05:30 crc kubenswrapper[4809]: I1124 07:05:30.238525 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5xvl8_3c932894-56b4-4303-8176-d12b12686b09/kube-multus/2.log" Nov 24 07:05:30 crc kubenswrapper[4809]: I1124 07:05:30.239077 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5xvl8" event={"ID":"3c932894-56b4-4303-8176-d12b12686b09","Type":"ContainerStarted","Data":"7b2d6da4cb8da00eaa9a608096a719f8bff1690346bab27fc3271d4509140b76"} Nov 24 07:05:31 crc kubenswrapper[4809]: I1124 07:05:31.521609 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zrmrp" Nov 24 07:05:44 crc kubenswrapper[4809]: I1124 07:05:44.501734 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv"] Nov 24 07:05:44 crc kubenswrapper[4809]: I1124 07:05:44.504875 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv"] Nov 24 07:05:44 crc kubenswrapper[4809]: I1124 07:05:44.505071 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv" Nov 24 07:05:44 crc kubenswrapper[4809]: I1124 07:05:44.508472 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 07:05:44 crc kubenswrapper[4809]: I1124 07:05:44.576309 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4bac6145-551f-4994-83ea-86181d6719ae-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv\" (UID: \"4bac6145-551f-4994-83ea-86181d6719ae\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv" Nov 24 07:05:44 crc kubenswrapper[4809]: I1124 07:05:44.576440 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4bac6145-551f-4994-83ea-86181d6719ae-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv\" (UID: \"4bac6145-551f-4994-83ea-86181d6719ae\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv" Nov 24 07:05:44 crc kubenswrapper[4809]: I1124 07:05:44.576597 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n8mg\" (UniqueName: \"kubernetes.io/projected/4bac6145-551f-4994-83ea-86181d6719ae-kube-api-access-9n8mg\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv\" (UID: \"4bac6145-551f-4994-83ea-86181d6719ae\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv" Nov 24 07:05:44 crc kubenswrapper[4809]: I1124 07:05:44.678073 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4bac6145-551f-4994-83ea-86181d6719ae-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv\" (UID: \"4bac6145-551f-4994-83ea-86181d6719ae\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv" Nov 24 07:05:44 crc kubenswrapper[4809]: I1124 07:05:44.678189 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n8mg\" (UniqueName: \"kubernetes.io/projected/4bac6145-551f-4994-83ea-86181d6719ae-kube-api-access-9n8mg\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv\" (UID: \"4bac6145-551f-4994-83ea-86181d6719ae\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv" Nov 24 07:05:44 crc kubenswrapper[4809]: I1124 07:05:44.678243 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4bac6145-551f-4994-83ea-86181d6719ae-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv\" (UID: \"4bac6145-551f-4994-83ea-86181d6719ae\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv" Nov 24 07:05:44 crc kubenswrapper[4809]: I1124 07:05:44.679042 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4bac6145-551f-4994-83ea-86181d6719ae-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv\" (UID: \"4bac6145-551f-4994-83ea-86181d6719ae\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv" Nov 24 07:05:44 crc kubenswrapper[4809]: I1124 07:05:44.679406 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4bac6145-551f-4994-83ea-86181d6719ae-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv\" (UID: \"4bac6145-551f-4994-83ea-86181d6719ae\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv" Nov 24 07:05:44 crc kubenswrapper[4809]: I1124 07:05:44.715805 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n8mg\" (UniqueName: \"kubernetes.io/projected/4bac6145-551f-4994-83ea-86181d6719ae-kube-api-access-9n8mg\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv\" (UID: \"4bac6145-551f-4994-83ea-86181d6719ae\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv" Nov 24 07:05:44 crc kubenswrapper[4809]: I1124 07:05:44.828297 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv" Nov 24 07:05:45 crc kubenswrapper[4809]: I1124 07:05:45.968609 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv"] Nov 24 07:05:45 crc kubenswrapper[4809]: W1124 07:05:45.978169 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4bac6145_551f_4994_83ea_86181d6719ae.slice/crio-374cab90ebf29e3b49f44383435ceeff75d0dc839f82140ec1f3f41eb0f34a7b WatchSource:0}: Error finding container 374cab90ebf29e3b49f44383435ceeff75d0dc839f82140ec1f3f41eb0f34a7b: Status 404 returned error can't find the container with id 374cab90ebf29e3b49f44383435ceeff75d0dc839f82140ec1f3f41eb0f34a7b Nov 24 07:05:46 crc kubenswrapper[4809]: I1124 07:05:46.868115 4809 generic.go:334] "Generic (PLEG): container finished" podID="4bac6145-551f-4994-83ea-86181d6719ae" containerID="e4b04e3423cadafc61a80c464b961d1c79174c32d424fcf33bd277ce75d2ddcd" exitCode=0 Nov 24 07:05:46 crc kubenswrapper[4809]: I1124 07:05:46.868514 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv" event={"ID":"4bac6145-551f-4994-83ea-86181d6719ae","Type":"ContainerDied","Data":"e4b04e3423cadafc61a80c464b961d1c79174c32d424fcf33bd277ce75d2ddcd"} Nov 24 07:05:46 crc kubenswrapper[4809]: I1124 07:05:46.870545 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv" event={"ID":"4bac6145-551f-4994-83ea-86181d6719ae","Type":"ContainerStarted","Data":"374cab90ebf29e3b49f44383435ceeff75d0dc839f82140ec1f3f41eb0f34a7b"} Nov 24 07:05:48 crc kubenswrapper[4809]: I1124 07:05:48.888137 4809 generic.go:334] "Generic (PLEG): container finished" podID="4bac6145-551f-4994-83ea-86181d6719ae" containerID="859a7e6baad82b6113682d6cf068c351dc55f5125dd4166febde8ddab7a1ccfb" exitCode=0 Nov 24 07:05:48 crc kubenswrapper[4809]: I1124 07:05:48.888195 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv" event={"ID":"4bac6145-551f-4994-83ea-86181d6719ae","Type":"ContainerDied","Data":"859a7e6baad82b6113682d6cf068c351dc55f5125dd4166febde8ddab7a1ccfb"} Nov 24 07:05:49 crc kubenswrapper[4809]: I1124 07:05:49.897946 4809 generic.go:334] "Generic (PLEG): container finished" podID="4bac6145-551f-4994-83ea-86181d6719ae" containerID="7f2dc1fd3ae17c3fe890bc2cea0a27ccc64c010028af7766edf87c06a433bffb" exitCode=0 Nov 24 07:05:49 crc kubenswrapper[4809]: I1124 07:05:49.898027 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv" event={"ID":"4bac6145-551f-4994-83ea-86181d6719ae","Type":"ContainerDied","Data":"7f2dc1fd3ae17c3fe890bc2cea0a27ccc64c010028af7766edf87c06a433bffb"} Nov 24 07:05:51 crc kubenswrapper[4809]: I1124 07:05:51.242900 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv" Nov 24 07:05:51 crc kubenswrapper[4809]: I1124 07:05:51.429892 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4bac6145-551f-4994-83ea-86181d6719ae-util\") pod \"4bac6145-551f-4994-83ea-86181d6719ae\" (UID: \"4bac6145-551f-4994-83ea-86181d6719ae\") " Nov 24 07:05:51 crc kubenswrapper[4809]: I1124 07:05:51.430101 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n8mg\" (UniqueName: \"kubernetes.io/projected/4bac6145-551f-4994-83ea-86181d6719ae-kube-api-access-9n8mg\") pod \"4bac6145-551f-4994-83ea-86181d6719ae\" (UID: \"4bac6145-551f-4994-83ea-86181d6719ae\") " Nov 24 07:05:51 crc kubenswrapper[4809]: I1124 07:05:51.430293 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4bac6145-551f-4994-83ea-86181d6719ae-bundle\") pod \"4bac6145-551f-4994-83ea-86181d6719ae\" (UID: \"4bac6145-551f-4994-83ea-86181d6719ae\") " Nov 24 07:05:51 crc kubenswrapper[4809]: I1124 07:05:51.431417 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bac6145-551f-4994-83ea-86181d6719ae-bundle" (OuterVolumeSpecName: "bundle") pod "4bac6145-551f-4994-83ea-86181d6719ae" (UID: "4bac6145-551f-4994-83ea-86181d6719ae"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:05:51 crc kubenswrapper[4809]: I1124 07:05:51.442462 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bac6145-551f-4994-83ea-86181d6719ae-kube-api-access-9n8mg" (OuterVolumeSpecName: "kube-api-access-9n8mg") pod "4bac6145-551f-4994-83ea-86181d6719ae" (UID: "4bac6145-551f-4994-83ea-86181d6719ae"). InnerVolumeSpecName "kube-api-access-9n8mg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:05:51 crc kubenswrapper[4809]: I1124 07:05:51.462071 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bac6145-551f-4994-83ea-86181d6719ae-util" (OuterVolumeSpecName: "util") pod "4bac6145-551f-4994-83ea-86181d6719ae" (UID: "4bac6145-551f-4994-83ea-86181d6719ae"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:05:51 crc kubenswrapper[4809]: I1124 07:05:51.532090 4809 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4bac6145-551f-4994-83ea-86181d6719ae-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:51 crc kubenswrapper[4809]: I1124 07:05:51.532162 4809 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4bac6145-551f-4994-83ea-86181d6719ae-util\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:51 crc kubenswrapper[4809]: I1124 07:05:51.532196 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n8mg\" (UniqueName: \"kubernetes.io/projected/4bac6145-551f-4994-83ea-86181d6719ae-kube-api-access-9n8mg\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:51 crc kubenswrapper[4809]: I1124 07:05:51.915028 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv" event={"ID":"4bac6145-551f-4994-83ea-86181d6719ae","Type":"ContainerDied","Data":"374cab90ebf29e3b49f44383435ceeff75d0dc839f82140ec1f3f41eb0f34a7b"} Nov 24 07:05:51 crc kubenswrapper[4809]: I1124 07:05:51.915314 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="374cab90ebf29e3b49f44383435ceeff75d0dc839f82140ec1f3f41eb0f34a7b" Nov 24 07:05:51 crc kubenswrapper[4809]: I1124 07:05:51.915152 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv" Nov 24 07:05:56 crc kubenswrapper[4809]: I1124 07:05:56.080439 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-hbbwq"] Nov 24 07:05:56 crc kubenswrapper[4809]: E1124 07:05:56.082334 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bac6145-551f-4994-83ea-86181d6719ae" containerName="extract" Nov 24 07:05:56 crc kubenswrapper[4809]: I1124 07:05:56.082446 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bac6145-551f-4994-83ea-86181d6719ae" containerName="extract" Nov 24 07:05:56 crc kubenswrapper[4809]: E1124 07:05:56.082554 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bac6145-551f-4994-83ea-86181d6719ae" containerName="pull" Nov 24 07:05:56 crc kubenswrapper[4809]: I1124 07:05:56.082637 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bac6145-551f-4994-83ea-86181d6719ae" containerName="pull" Nov 24 07:05:56 crc kubenswrapper[4809]: E1124 07:05:56.082709 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bac6145-551f-4994-83ea-86181d6719ae" containerName="util" Nov 24 07:05:56 crc kubenswrapper[4809]: I1124 07:05:56.082773 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bac6145-551f-4994-83ea-86181d6719ae" containerName="util" Nov 24 07:05:56 crc kubenswrapper[4809]: I1124 07:05:56.083016 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bac6145-551f-4994-83ea-86181d6719ae" containerName="extract" Nov 24 07:05:56 crc kubenswrapper[4809]: I1124 07:05:56.083550 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-hbbwq" Nov 24 07:05:56 crc kubenswrapper[4809]: I1124 07:05:56.085528 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 24 07:05:56 crc kubenswrapper[4809]: I1124 07:05:56.085704 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-jwp6d" Nov 24 07:05:56 crc kubenswrapper[4809]: I1124 07:05:56.086104 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 24 07:05:56 crc kubenswrapper[4809]: I1124 07:05:56.103730 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-hbbwq"] Nov 24 07:05:56 crc kubenswrapper[4809]: I1124 07:05:56.196187 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjgz4\" (UniqueName: \"kubernetes.io/projected/dec4efb0-1f05-416f-8e23-773aaab020df-kube-api-access-mjgz4\") pod \"nmstate-operator-557fdffb88-hbbwq\" (UID: \"dec4efb0-1f05-416f-8e23-773aaab020df\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-hbbwq" Nov 24 07:05:56 crc kubenswrapper[4809]: I1124 07:05:56.297438 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjgz4\" (UniqueName: \"kubernetes.io/projected/dec4efb0-1f05-416f-8e23-773aaab020df-kube-api-access-mjgz4\") pod \"nmstate-operator-557fdffb88-hbbwq\" (UID: \"dec4efb0-1f05-416f-8e23-773aaab020df\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-hbbwq" Nov 24 07:05:56 crc kubenswrapper[4809]: I1124 07:05:56.334477 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjgz4\" (UniqueName: \"kubernetes.io/projected/dec4efb0-1f05-416f-8e23-773aaab020df-kube-api-access-mjgz4\") pod \"nmstate-operator-557fdffb88-hbbwq\" (UID: \"dec4efb0-1f05-416f-8e23-773aaab020df\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-hbbwq" Nov 24 07:05:56 crc kubenswrapper[4809]: I1124 07:05:56.400823 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-hbbwq" Nov 24 07:05:56 crc kubenswrapper[4809]: I1124 07:05:56.962361 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-hbbwq"] Nov 24 07:05:57 crc kubenswrapper[4809]: I1124 07:05:57.956793 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-hbbwq" event={"ID":"dec4efb0-1f05-416f-8e23-773aaab020df","Type":"ContainerStarted","Data":"3f8b64125ebe2b5e454aab324903a580b177b9d56a4fb5e39f97c34041ab3a0e"} Nov 24 07:05:58 crc kubenswrapper[4809]: I1124 07:05:58.966552 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-hbbwq" event={"ID":"dec4efb0-1f05-416f-8e23-773aaab020df","Type":"ContainerStarted","Data":"1d0e92e97983a61160bd3beb040cb17e779a3547f771514563bf907b25ffc51a"} Nov 24 07:05:58 crc kubenswrapper[4809]: I1124 07:05:58.988872 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-hbbwq" podStartSLOduration=1.365793407 podStartE2EDuration="2.988845601s" podCreationTimestamp="2025-11-24 07:05:56 +0000 UTC" firstStartedPulling="2025-11-24 07:05:56.973125654 +0000 UTC m=+656.873717349" lastFinishedPulling="2025-11-24 07:05:58.596177838 +0000 UTC m=+658.496769543" observedRunningTime="2025-11-24 07:05:58.982435297 +0000 UTC m=+658.883027032" watchObservedRunningTime="2025-11-24 07:05:58.988845601 +0000 UTC m=+658.889437346" Nov 24 07:06:04 crc kubenswrapper[4809]: I1124 07:06:04.884585 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-782jg"] Nov 24 07:06:04 crc kubenswrapper[4809]: I1124 07:06:04.886177 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-782jg" Nov 24 07:06:04 crc kubenswrapper[4809]: I1124 07:06:04.888683 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-67vjf" Nov 24 07:06:04 crc kubenswrapper[4809]: I1124 07:06:04.901760 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-7fhk7"] Nov 24 07:06:04 crc kubenswrapper[4809]: I1124 07:06:04.902433 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-7fhk7" Nov 24 07:06:04 crc kubenswrapper[4809]: I1124 07:06:04.904584 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 24 07:06:04 crc kubenswrapper[4809]: I1124 07:06:04.905984 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-782jg"] Nov 24 07:06:04 crc kubenswrapper[4809]: I1124 07:06:04.925026 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-7fhk7"] Nov 24 07:06:04 crc kubenswrapper[4809]: I1124 07:06:04.927942 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-5dl97"] Nov 24 07:06:04 crc kubenswrapper[4809]: I1124 07:06:04.928691 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-5dl97" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.018506 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-8fgt8"] Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.019334 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-8fgt8" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.022757 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.022795 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-gd68t" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.022858 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.028446 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-8fgt8"] Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.035392 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6cw5\" (UniqueName: \"kubernetes.io/projected/c00579c8-50fc-4b4e-8ddf-79ccdc0647ec-kube-api-access-c6cw5\") pod \"nmstate-metrics-5dcf9c57c5-782jg\" (UID: \"c00579c8-50fc-4b4e-8ddf-79ccdc0647ec\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-782jg" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.035436 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ef4d3282-be25-4395-899d-92b0fc8a0a40-dbus-socket\") pod \"nmstate-handler-5dl97\" (UID: \"ef4d3282-be25-4395-899d-92b0fc8a0a40\") " pod="openshift-nmstate/nmstate-handler-5dl97" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.035461 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgl4p\" (UniqueName: \"kubernetes.io/projected/5c12932d-5d71-4a4f-9dc5-bbf20584f029-kube-api-access-bgl4p\") pod \"nmstate-webhook-6b89b748d8-7fhk7\" (UID: \"5c12932d-5d71-4a4f-9dc5-bbf20584f029\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-7fhk7" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.035478 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ef4d3282-be25-4395-899d-92b0fc8a0a40-ovs-socket\") pod \"nmstate-handler-5dl97\" (UID: \"ef4d3282-be25-4395-899d-92b0fc8a0a40\") " pod="openshift-nmstate/nmstate-handler-5dl97" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.035496 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ef4d3282-be25-4395-899d-92b0fc8a0a40-nmstate-lock\") pod \"nmstate-handler-5dl97\" (UID: \"ef4d3282-be25-4395-899d-92b0fc8a0a40\") " pod="openshift-nmstate/nmstate-handler-5dl97" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.035548 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/5c12932d-5d71-4a4f-9dc5-bbf20584f029-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-7fhk7\" (UID: \"5c12932d-5d71-4a4f-9dc5-bbf20584f029\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-7fhk7" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.035572 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcxzp\" (UniqueName: \"kubernetes.io/projected/ef4d3282-be25-4395-899d-92b0fc8a0a40-kube-api-access-vcxzp\") pod \"nmstate-handler-5dl97\" (UID: \"ef4d3282-be25-4395-899d-92b0fc8a0a40\") " pod="openshift-nmstate/nmstate-handler-5dl97" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.136917 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6cw5\" (UniqueName: \"kubernetes.io/projected/c00579c8-50fc-4b4e-8ddf-79ccdc0647ec-kube-api-access-c6cw5\") pod \"nmstate-metrics-5dcf9c57c5-782jg\" (UID: \"c00579c8-50fc-4b4e-8ddf-79ccdc0647ec\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-782jg" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.137248 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ef4d3282-be25-4395-899d-92b0fc8a0a40-dbus-socket\") pod \"nmstate-handler-5dl97\" (UID: \"ef4d3282-be25-4395-899d-92b0fc8a0a40\") " pod="openshift-nmstate/nmstate-handler-5dl97" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.137273 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgl4p\" (UniqueName: \"kubernetes.io/projected/5c12932d-5d71-4a4f-9dc5-bbf20584f029-kube-api-access-bgl4p\") pod \"nmstate-webhook-6b89b748d8-7fhk7\" (UID: \"5c12932d-5d71-4a4f-9dc5-bbf20584f029\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-7fhk7" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.137296 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ef4d3282-be25-4395-899d-92b0fc8a0a40-ovs-socket\") pod \"nmstate-handler-5dl97\" (UID: \"ef4d3282-be25-4395-899d-92b0fc8a0a40\") " pod="openshift-nmstate/nmstate-handler-5dl97" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.137313 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ef4d3282-be25-4395-899d-92b0fc8a0a40-nmstate-lock\") pod \"nmstate-handler-5dl97\" (UID: \"ef4d3282-be25-4395-899d-92b0fc8a0a40\") " pod="openshift-nmstate/nmstate-handler-5dl97" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.137348 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7265684f-a24a-40d3-8eb5-eb2da39cb870-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-8fgt8\" (UID: \"7265684f-a24a-40d3-8eb5-eb2da39cb870\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-8fgt8" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.137393 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/5c12932d-5d71-4a4f-9dc5-bbf20584f029-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-7fhk7\" (UID: \"5c12932d-5d71-4a4f-9dc5-bbf20584f029\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-7fhk7" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.137398 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ef4d3282-be25-4395-899d-92b0fc8a0a40-ovs-socket\") pod \"nmstate-handler-5dl97\" (UID: \"ef4d3282-be25-4395-899d-92b0fc8a0a40\") " pod="openshift-nmstate/nmstate-handler-5dl97" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.137412 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkzgl\" (UniqueName: \"kubernetes.io/projected/7265684f-a24a-40d3-8eb5-eb2da39cb870-kube-api-access-vkzgl\") pod \"nmstate-console-plugin-5874bd7bc5-8fgt8\" (UID: \"7265684f-a24a-40d3-8eb5-eb2da39cb870\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-8fgt8" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.137458 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ef4d3282-be25-4395-899d-92b0fc8a0a40-nmstate-lock\") pod \"nmstate-handler-5dl97\" (UID: \"ef4d3282-be25-4395-899d-92b0fc8a0a40\") " pod="openshift-nmstate/nmstate-handler-5dl97" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.137501 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcxzp\" (UniqueName: \"kubernetes.io/projected/ef4d3282-be25-4395-899d-92b0fc8a0a40-kube-api-access-vcxzp\") pod \"nmstate-handler-5dl97\" (UID: \"ef4d3282-be25-4395-899d-92b0fc8a0a40\") " pod="openshift-nmstate/nmstate-handler-5dl97" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.137563 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7265684f-a24a-40d3-8eb5-eb2da39cb870-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-8fgt8\" (UID: \"7265684f-a24a-40d3-8eb5-eb2da39cb870\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-8fgt8" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.137627 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ef4d3282-be25-4395-899d-92b0fc8a0a40-dbus-socket\") pod \"nmstate-handler-5dl97\" (UID: \"ef4d3282-be25-4395-899d-92b0fc8a0a40\") " pod="openshift-nmstate/nmstate-handler-5dl97" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.143808 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/5c12932d-5d71-4a4f-9dc5-bbf20584f029-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-7fhk7\" (UID: \"5c12932d-5d71-4a4f-9dc5-bbf20584f029\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-7fhk7" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.152912 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6cw5\" (UniqueName: \"kubernetes.io/projected/c00579c8-50fc-4b4e-8ddf-79ccdc0647ec-kube-api-access-c6cw5\") pod \"nmstate-metrics-5dcf9c57c5-782jg\" (UID: \"c00579c8-50fc-4b4e-8ddf-79ccdc0647ec\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-782jg" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.157640 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgl4p\" (UniqueName: \"kubernetes.io/projected/5c12932d-5d71-4a4f-9dc5-bbf20584f029-kube-api-access-bgl4p\") pod \"nmstate-webhook-6b89b748d8-7fhk7\" (UID: \"5c12932d-5d71-4a4f-9dc5-bbf20584f029\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-7fhk7" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.159657 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcxzp\" (UniqueName: \"kubernetes.io/projected/ef4d3282-be25-4395-899d-92b0fc8a0a40-kube-api-access-vcxzp\") pod \"nmstate-handler-5dl97\" (UID: \"ef4d3282-be25-4395-899d-92b0fc8a0a40\") " pod="openshift-nmstate/nmstate-handler-5dl97" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.206108 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-782jg" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.212732 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-66bb954594-cbvcz"] Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.213514 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.218431 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-7fhk7" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.227164 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-66bb954594-cbvcz"] Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.238830 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkzgl\" (UniqueName: \"kubernetes.io/projected/7265684f-a24a-40d3-8eb5-eb2da39cb870-kube-api-access-vkzgl\") pod \"nmstate-console-plugin-5874bd7bc5-8fgt8\" (UID: \"7265684f-a24a-40d3-8eb5-eb2da39cb870\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-8fgt8" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.238885 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7265684f-a24a-40d3-8eb5-eb2da39cb870-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-8fgt8\" (UID: \"7265684f-a24a-40d3-8eb5-eb2da39cb870\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-8fgt8" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.238943 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7265684f-a24a-40d3-8eb5-eb2da39cb870-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-8fgt8\" (UID: \"7265684f-a24a-40d3-8eb5-eb2da39cb870\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-8fgt8" Nov 24 07:06:05 crc kubenswrapper[4809]: E1124 07:06:05.239066 4809 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Nov 24 07:06:05 crc kubenswrapper[4809]: E1124 07:06:05.239141 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7265684f-a24a-40d3-8eb5-eb2da39cb870-plugin-serving-cert podName:7265684f-a24a-40d3-8eb5-eb2da39cb870 nodeName:}" failed. No retries permitted until 2025-11-24 07:06:05.739117887 +0000 UTC m=+665.639709592 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/7265684f-a24a-40d3-8eb5-eb2da39cb870-plugin-serving-cert") pod "nmstate-console-plugin-5874bd7bc5-8fgt8" (UID: "7265684f-a24a-40d3-8eb5-eb2da39cb870") : secret "plugin-serving-cert" not found Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.239900 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7265684f-a24a-40d3-8eb5-eb2da39cb870-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-8fgt8\" (UID: \"7265684f-a24a-40d3-8eb5-eb2da39cb870\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-8fgt8" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.243590 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-5dl97" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.260273 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkzgl\" (UniqueName: \"kubernetes.io/projected/7265684f-a24a-40d3-8eb5-eb2da39cb870-kube-api-access-vkzgl\") pod \"nmstate-console-plugin-5874bd7bc5-8fgt8\" (UID: \"7265684f-a24a-40d3-8eb5-eb2da39cb870\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-8fgt8" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.340923 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ck5f9\" (UniqueName: \"kubernetes.io/projected/52ce38ba-4c75-4f7d-b41b-65a692a94bdb-kube-api-access-ck5f9\") pod \"console-66bb954594-cbvcz\" (UID: \"52ce38ba-4c75-4f7d-b41b-65a692a94bdb\") " pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.341245 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/52ce38ba-4c75-4f7d-b41b-65a692a94bdb-trusted-ca-bundle\") pod \"console-66bb954594-cbvcz\" (UID: \"52ce38ba-4c75-4f7d-b41b-65a692a94bdb\") " pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.341267 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/52ce38ba-4c75-4f7d-b41b-65a692a94bdb-console-oauth-config\") pod \"console-66bb954594-cbvcz\" (UID: \"52ce38ba-4c75-4f7d-b41b-65a692a94bdb\") " pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.341300 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/52ce38ba-4c75-4f7d-b41b-65a692a94bdb-oauth-serving-cert\") pod \"console-66bb954594-cbvcz\" (UID: \"52ce38ba-4c75-4f7d-b41b-65a692a94bdb\") " pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.341327 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/52ce38ba-4c75-4f7d-b41b-65a692a94bdb-service-ca\") pod \"console-66bb954594-cbvcz\" (UID: \"52ce38ba-4c75-4f7d-b41b-65a692a94bdb\") " pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.341345 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/52ce38ba-4c75-4f7d-b41b-65a692a94bdb-console-serving-cert\") pod \"console-66bb954594-cbvcz\" (UID: \"52ce38ba-4c75-4f7d-b41b-65a692a94bdb\") " pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.341361 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/52ce38ba-4c75-4f7d-b41b-65a692a94bdb-console-config\") pod \"console-66bb954594-cbvcz\" (UID: \"52ce38ba-4c75-4f7d-b41b-65a692a94bdb\") " pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.445179 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ck5f9\" (UniqueName: \"kubernetes.io/projected/52ce38ba-4c75-4f7d-b41b-65a692a94bdb-kube-api-access-ck5f9\") pod \"console-66bb954594-cbvcz\" (UID: \"52ce38ba-4c75-4f7d-b41b-65a692a94bdb\") " pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.445237 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/52ce38ba-4c75-4f7d-b41b-65a692a94bdb-trusted-ca-bundle\") pod \"console-66bb954594-cbvcz\" (UID: \"52ce38ba-4c75-4f7d-b41b-65a692a94bdb\") " pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.445266 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/52ce38ba-4c75-4f7d-b41b-65a692a94bdb-console-oauth-config\") pod \"console-66bb954594-cbvcz\" (UID: \"52ce38ba-4c75-4f7d-b41b-65a692a94bdb\") " pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.445300 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/52ce38ba-4c75-4f7d-b41b-65a692a94bdb-oauth-serving-cert\") pod \"console-66bb954594-cbvcz\" (UID: \"52ce38ba-4c75-4f7d-b41b-65a692a94bdb\") " pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.445335 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/52ce38ba-4c75-4f7d-b41b-65a692a94bdb-service-ca\") pod \"console-66bb954594-cbvcz\" (UID: \"52ce38ba-4c75-4f7d-b41b-65a692a94bdb\") " pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.445358 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/52ce38ba-4c75-4f7d-b41b-65a692a94bdb-console-serving-cert\") pod \"console-66bb954594-cbvcz\" (UID: \"52ce38ba-4c75-4f7d-b41b-65a692a94bdb\") " pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.445395 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/52ce38ba-4c75-4f7d-b41b-65a692a94bdb-console-config\") pod \"console-66bb954594-cbvcz\" (UID: \"52ce38ba-4c75-4f7d-b41b-65a692a94bdb\") " pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.446577 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/52ce38ba-4c75-4f7d-b41b-65a692a94bdb-console-config\") pod \"console-66bb954594-cbvcz\" (UID: \"52ce38ba-4c75-4f7d-b41b-65a692a94bdb\") " pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.446645 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/52ce38ba-4c75-4f7d-b41b-65a692a94bdb-oauth-serving-cert\") pod \"console-66bb954594-cbvcz\" (UID: \"52ce38ba-4c75-4f7d-b41b-65a692a94bdb\") " pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.446691 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/52ce38ba-4c75-4f7d-b41b-65a692a94bdb-service-ca\") pod \"console-66bb954594-cbvcz\" (UID: \"52ce38ba-4c75-4f7d-b41b-65a692a94bdb\") " pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.450713 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/52ce38ba-4c75-4f7d-b41b-65a692a94bdb-console-serving-cert\") pod \"console-66bb954594-cbvcz\" (UID: \"52ce38ba-4c75-4f7d-b41b-65a692a94bdb\") " pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.454102 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/52ce38ba-4c75-4f7d-b41b-65a692a94bdb-console-oauth-config\") pod \"console-66bb954594-cbvcz\" (UID: \"52ce38ba-4c75-4f7d-b41b-65a692a94bdb\") " pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.455240 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/52ce38ba-4c75-4f7d-b41b-65a692a94bdb-trusted-ca-bundle\") pod \"console-66bb954594-cbvcz\" (UID: \"52ce38ba-4c75-4f7d-b41b-65a692a94bdb\") " pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.461173 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-7fhk7"] Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.466212 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ck5f9\" (UniqueName: \"kubernetes.io/projected/52ce38ba-4c75-4f7d-b41b-65a692a94bdb-kube-api-access-ck5f9\") pod \"console-66bb954594-cbvcz\" (UID: \"52ce38ba-4c75-4f7d-b41b-65a692a94bdb\") " pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.620794 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.649677 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-782jg"] Nov 24 07:06:05 crc kubenswrapper[4809]: W1124 07:06:05.654881 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc00579c8_50fc_4b4e_8ddf_79ccdc0647ec.slice/crio-cea36d8628d5e44e77a222d929e7f623a726fdc1ca0d2e0df6037016c00d56cd WatchSource:0}: Error finding container cea36d8628d5e44e77a222d929e7f623a726fdc1ca0d2e0df6037016c00d56cd: Status 404 returned error can't find the container with id cea36d8628d5e44e77a222d929e7f623a726fdc1ca0d2e0df6037016c00d56cd Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.749846 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7265684f-a24a-40d3-8eb5-eb2da39cb870-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-8fgt8\" (UID: \"7265684f-a24a-40d3-8eb5-eb2da39cb870\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-8fgt8" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.753000 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7265684f-a24a-40d3-8eb5-eb2da39cb870-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-8fgt8\" (UID: \"7265684f-a24a-40d3-8eb5-eb2da39cb870\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-8fgt8" Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.787581 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-66bb954594-cbvcz"] Nov 24 07:06:05 crc kubenswrapper[4809]: I1124 07:06:05.937906 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-8fgt8" Nov 24 07:06:06 crc kubenswrapper[4809]: I1124 07:06:06.018900 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-7fhk7" event={"ID":"5c12932d-5d71-4a4f-9dc5-bbf20584f029","Type":"ContainerStarted","Data":"2f92d703a5612ec10d83b1d8fed6366bf3f37e914bf91add573578de07c78e63"} Nov 24 07:06:06 crc kubenswrapper[4809]: I1124 07:06:06.020066 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-66bb954594-cbvcz" event={"ID":"52ce38ba-4c75-4f7d-b41b-65a692a94bdb","Type":"ContainerStarted","Data":"6b6cf416d2364fde3c753f828a5082ac9eb019cba7c5d651547abb15c5eb274e"} Nov 24 07:06:06 crc kubenswrapper[4809]: I1124 07:06:06.020104 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-66bb954594-cbvcz" event={"ID":"52ce38ba-4c75-4f7d-b41b-65a692a94bdb","Type":"ContainerStarted","Data":"0946bd358fb9944a482aa4e72b97b506b51531bbc041d2b4d2120a2337be0700"} Nov 24 07:06:06 crc kubenswrapper[4809]: I1124 07:06:06.021635 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-5dl97" event={"ID":"ef4d3282-be25-4395-899d-92b0fc8a0a40","Type":"ContainerStarted","Data":"52e81b5ddf2ff6daadd714ba81eda8c0bea195523139815e7195a9c6e95a6818"} Nov 24 07:06:06 crc kubenswrapper[4809]: I1124 07:06:06.023050 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-782jg" event={"ID":"c00579c8-50fc-4b4e-8ddf-79ccdc0647ec","Type":"ContainerStarted","Data":"cea36d8628d5e44e77a222d929e7f623a726fdc1ca0d2e0df6037016c00d56cd"} Nov 24 07:06:06 crc kubenswrapper[4809]: I1124 07:06:06.047151 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-66bb954594-cbvcz" podStartSLOduration=1.047131648 podStartE2EDuration="1.047131648s" podCreationTimestamp="2025-11-24 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:06:06.044715363 +0000 UTC m=+665.945307088" watchObservedRunningTime="2025-11-24 07:06:06.047131648 +0000 UTC m=+665.947723353" Nov 24 07:06:06 crc kubenswrapper[4809]: I1124 07:06:06.138070 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-8fgt8"] Nov 24 07:06:06 crc kubenswrapper[4809]: W1124 07:06:06.142553 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7265684f_a24a_40d3_8eb5_eb2da39cb870.slice/crio-0a841ef3092776062d44e43faa621a7f9479c467b295d8dc54a6b6eb7a2bf714 WatchSource:0}: Error finding container 0a841ef3092776062d44e43faa621a7f9479c467b295d8dc54a6b6eb7a2bf714: Status 404 returned error can't find the container with id 0a841ef3092776062d44e43faa621a7f9479c467b295d8dc54a6b6eb7a2bf714 Nov 24 07:06:07 crc kubenswrapper[4809]: I1124 07:06:07.030699 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-8fgt8" event={"ID":"7265684f-a24a-40d3-8eb5-eb2da39cb870","Type":"ContainerStarted","Data":"0a841ef3092776062d44e43faa621a7f9479c467b295d8dc54a6b6eb7a2bf714"} Nov 24 07:06:08 crc kubenswrapper[4809]: I1124 07:06:08.037448 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-782jg" event={"ID":"c00579c8-50fc-4b4e-8ddf-79ccdc0647ec","Type":"ContainerStarted","Data":"b356f9e4397aa6ad9aa157b59f9300fe95124a98590a58e8280444e97be7dc7b"} Nov 24 07:06:08 crc kubenswrapper[4809]: I1124 07:06:08.038911 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-7fhk7" event={"ID":"5c12932d-5d71-4a4f-9dc5-bbf20584f029","Type":"ContainerStarted","Data":"6c97e4f1822d5e8ad12b86eae98accd60596df9061967a34fba9cd8e4ec6ab24"} Nov 24 07:06:08 crc kubenswrapper[4809]: I1124 07:06:08.039134 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-7fhk7" Nov 24 07:06:08 crc kubenswrapper[4809]: I1124 07:06:08.041177 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-5dl97" event={"ID":"ef4d3282-be25-4395-899d-92b0fc8a0a40","Type":"ContainerStarted","Data":"0f296bf3fbe0a0c77045fbcf54acd44db0c6168c49bb9040af85c4924571a0ea"} Nov 24 07:06:08 crc kubenswrapper[4809]: I1124 07:06:08.041283 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-5dl97" Nov 24 07:06:08 crc kubenswrapper[4809]: I1124 07:06:08.057098 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-7fhk7" podStartSLOduration=1.979809223 podStartE2EDuration="4.057078428s" podCreationTimestamp="2025-11-24 07:06:04 +0000 UTC" firstStartedPulling="2025-11-24 07:06:05.470164826 +0000 UTC m=+665.370756531" lastFinishedPulling="2025-11-24 07:06:07.547434001 +0000 UTC m=+667.448025736" observedRunningTime="2025-11-24 07:06:08.056126292 +0000 UTC m=+667.956718017" watchObservedRunningTime="2025-11-24 07:06:08.057078428 +0000 UTC m=+667.957670133" Nov 24 07:06:08 crc kubenswrapper[4809]: I1124 07:06:08.070057 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-5dl97" podStartSLOduration=1.846142957 podStartE2EDuration="4.07003797s" podCreationTimestamp="2025-11-24 07:06:04 +0000 UTC" firstStartedPulling="2025-11-24 07:06:05.299190997 +0000 UTC m=+665.199782692" lastFinishedPulling="2025-11-24 07:06:07.523086 +0000 UTC m=+667.423677705" observedRunningTime="2025-11-24 07:06:08.069248728 +0000 UTC m=+667.969840443" watchObservedRunningTime="2025-11-24 07:06:08.07003797 +0000 UTC m=+667.970629675" Nov 24 07:06:09 crc kubenswrapper[4809]: I1124 07:06:09.049139 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-8fgt8" event={"ID":"7265684f-a24a-40d3-8eb5-eb2da39cb870","Type":"ContainerStarted","Data":"6a52c78055a34afea82bf3cbc8568204aa0b8d8a497e49fa0bc8683fc0314374"} Nov 24 07:06:09 crc kubenswrapper[4809]: I1124 07:06:09.064727 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-8fgt8" podStartSLOduration=1.613192056 podStartE2EDuration="4.064711735s" podCreationTimestamp="2025-11-24 07:06:05 +0000 UTC" firstStartedPulling="2025-11-24 07:06:06.145808146 +0000 UTC m=+666.046399851" lastFinishedPulling="2025-11-24 07:06:08.597327805 +0000 UTC m=+668.497919530" observedRunningTime="2025-11-24 07:06:09.064093757 +0000 UTC m=+668.964685472" watchObservedRunningTime="2025-11-24 07:06:09.064711735 +0000 UTC m=+668.965303440" Nov 24 07:06:10 crc kubenswrapper[4809]: I1124 07:06:10.056193 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-782jg" event={"ID":"c00579c8-50fc-4b4e-8ddf-79ccdc0647ec","Type":"ContainerStarted","Data":"16855643b588629dcc1d23ef951fff849b1d50a529bdd17956e78ab8aa908810"} Nov 24 07:06:15 crc kubenswrapper[4809]: I1124 07:06:15.278235 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-5dl97" Nov 24 07:06:15 crc kubenswrapper[4809]: I1124 07:06:15.307764 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-782jg" podStartSLOduration=7.272648283 podStartE2EDuration="11.307738805s" podCreationTimestamp="2025-11-24 07:06:04 +0000 UTC" firstStartedPulling="2025-11-24 07:06:05.657420155 +0000 UTC m=+665.558011860" lastFinishedPulling="2025-11-24 07:06:09.692510657 +0000 UTC m=+669.593102382" observedRunningTime="2025-11-24 07:06:10.080126542 +0000 UTC m=+669.980718297" watchObservedRunningTime="2025-11-24 07:06:15.307738805 +0000 UTC m=+675.208330540" Nov 24 07:06:15 crc kubenswrapper[4809]: I1124 07:06:15.621995 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:15 crc kubenswrapper[4809]: I1124 07:06:15.622070 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:15 crc kubenswrapper[4809]: I1124 07:06:15.630198 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:16 crc kubenswrapper[4809]: I1124 07:06:16.105626 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-66bb954594-cbvcz" Nov 24 07:06:16 crc kubenswrapper[4809]: I1124 07:06:16.177099 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-8r5vw"] Nov 24 07:06:25 crc kubenswrapper[4809]: I1124 07:06:25.225714 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-7fhk7" Nov 24 07:06:37 crc kubenswrapper[4809]: I1124 07:06:37.835622 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8"] Nov 24 07:06:37 crc kubenswrapper[4809]: I1124 07:06:37.837609 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8" Nov 24 07:06:37 crc kubenswrapper[4809]: I1124 07:06:37.840203 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 07:06:37 crc kubenswrapper[4809]: I1124 07:06:37.846364 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8"] Nov 24 07:06:38 crc kubenswrapper[4809]: I1124 07:06:38.034170 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7feb42ed-e415-44b8-a837-da74daf4aa52-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8\" (UID: \"7feb42ed-e415-44b8-a837-da74daf4aa52\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8" Nov 24 07:06:38 crc kubenswrapper[4809]: I1124 07:06:38.034218 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7feb42ed-e415-44b8-a837-da74daf4aa52-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8\" (UID: \"7feb42ed-e415-44b8-a837-da74daf4aa52\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8" Nov 24 07:06:38 crc kubenswrapper[4809]: I1124 07:06:38.034243 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw7mh\" (UniqueName: \"kubernetes.io/projected/7feb42ed-e415-44b8-a837-da74daf4aa52-kube-api-access-bw7mh\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8\" (UID: \"7feb42ed-e415-44b8-a837-da74daf4aa52\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8" Nov 24 07:06:38 crc kubenswrapper[4809]: I1124 07:06:38.136388 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7feb42ed-e415-44b8-a837-da74daf4aa52-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8\" (UID: \"7feb42ed-e415-44b8-a837-da74daf4aa52\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8" Nov 24 07:06:38 crc kubenswrapper[4809]: I1124 07:06:38.136518 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7feb42ed-e415-44b8-a837-da74daf4aa52-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8\" (UID: \"7feb42ed-e415-44b8-a837-da74daf4aa52\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8" Nov 24 07:06:38 crc kubenswrapper[4809]: I1124 07:06:38.136601 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw7mh\" (UniqueName: \"kubernetes.io/projected/7feb42ed-e415-44b8-a837-da74daf4aa52-kube-api-access-bw7mh\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8\" (UID: \"7feb42ed-e415-44b8-a837-da74daf4aa52\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8" Nov 24 07:06:38 crc kubenswrapper[4809]: I1124 07:06:38.137065 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7feb42ed-e415-44b8-a837-da74daf4aa52-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8\" (UID: \"7feb42ed-e415-44b8-a837-da74daf4aa52\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8" Nov 24 07:06:38 crc kubenswrapper[4809]: I1124 07:06:38.137094 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7feb42ed-e415-44b8-a837-da74daf4aa52-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8\" (UID: \"7feb42ed-e415-44b8-a837-da74daf4aa52\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8" Nov 24 07:06:38 crc kubenswrapper[4809]: I1124 07:06:38.161141 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw7mh\" (UniqueName: \"kubernetes.io/projected/7feb42ed-e415-44b8-a837-da74daf4aa52-kube-api-access-bw7mh\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8\" (UID: \"7feb42ed-e415-44b8-a837-da74daf4aa52\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8" Nov 24 07:06:38 crc kubenswrapper[4809]: I1124 07:06:38.453112 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8" Nov 24 07:06:38 crc kubenswrapper[4809]: I1124 07:06:38.620648 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8"] Nov 24 07:06:39 crc kubenswrapper[4809]: I1124 07:06:39.237696 4809 generic.go:334] "Generic (PLEG): container finished" podID="7feb42ed-e415-44b8-a837-da74daf4aa52" containerID="fee39fbcb602625d5763a0ade069a5b44b60e1f28ec1a2f2546dcf9f7776494e" exitCode=0 Nov 24 07:06:39 crc kubenswrapper[4809]: I1124 07:06:39.237992 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8" event={"ID":"7feb42ed-e415-44b8-a837-da74daf4aa52","Type":"ContainerDied","Data":"fee39fbcb602625d5763a0ade069a5b44b60e1f28ec1a2f2546dcf9f7776494e"} Nov 24 07:06:39 crc kubenswrapper[4809]: I1124 07:06:39.238022 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8" event={"ID":"7feb42ed-e415-44b8-a837-da74daf4aa52","Type":"ContainerStarted","Data":"b78533784f968363c5b7dce1f568865e6c8a44e54a8d07b64838f61514f80a1f"} Nov 24 07:06:40 crc kubenswrapper[4809]: I1124 07:06:40.249678 4809 generic.go:334] "Generic (PLEG): container finished" podID="7feb42ed-e415-44b8-a837-da74daf4aa52" containerID="caedf1d8c5b67ffd205bc2915333319ec8df29371c77ff4c394bed4643f5fb96" exitCode=0 Nov 24 07:06:40 crc kubenswrapper[4809]: I1124 07:06:40.249768 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8" event={"ID":"7feb42ed-e415-44b8-a837-da74daf4aa52","Type":"ContainerDied","Data":"caedf1d8c5b67ffd205bc2915333319ec8df29371c77ff4c394bed4643f5fb96"} Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.233686 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-8r5vw" podUID="2fe229a6-e9e2-40b3-a4d7-0780c4f5c884" containerName="console" containerID="cri-o://a167cf5d10e05e3dc032dd5a98a000f04efc29c0ee74ed9e382c4405089d041f" gracePeriod=15 Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.259406 4809 generic.go:334] "Generic (PLEG): container finished" podID="7feb42ed-e415-44b8-a837-da74daf4aa52" containerID="dc14df8241f0c02cd48bb2cb121dab1fb5dbf5884094f8d35011312f0a13784f" exitCode=0 Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.259451 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8" event={"ID":"7feb42ed-e415-44b8-a837-da74daf4aa52","Type":"ContainerDied","Data":"dc14df8241f0c02cd48bb2cb121dab1fb5dbf5884094f8d35011312f0a13784f"} Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.557708 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-8r5vw_2fe229a6-e9e2-40b3-a4d7-0780c4f5c884/console/0.log" Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.557792 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.682648 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-oauth-serving-cert\") pod \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.682853 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-console-serving-cert\") pod \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.682873 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-console-oauth-config\") pod \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.682940 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-service-ca\") pod \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.682984 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-console-config\") pod \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.683016 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-trusted-ca-bundle\") pod \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.683078 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bg99v\" (UniqueName: \"kubernetes.io/projected/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-kube-api-access-bg99v\") pod \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\" (UID: \"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884\") " Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.683780 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-service-ca" (OuterVolumeSpecName: "service-ca") pod "2fe229a6-e9e2-40b3-a4d7-0780c4f5c884" (UID: "2fe229a6-e9e2-40b3-a4d7-0780c4f5c884"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.684084 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "2fe229a6-e9e2-40b3-a4d7-0780c4f5c884" (UID: "2fe229a6-e9e2-40b3-a4d7-0780c4f5c884"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.684133 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-console-config" (OuterVolumeSpecName: "console-config") pod "2fe229a6-e9e2-40b3-a4d7-0780c4f5c884" (UID: "2fe229a6-e9e2-40b3-a4d7-0780c4f5c884"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.684086 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "2fe229a6-e9e2-40b3-a4d7-0780c4f5c884" (UID: "2fe229a6-e9e2-40b3-a4d7-0780c4f5c884"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.689638 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "2fe229a6-e9e2-40b3-a4d7-0780c4f5c884" (UID: "2fe229a6-e9e2-40b3-a4d7-0780c4f5c884"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.689639 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-kube-api-access-bg99v" (OuterVolumeSpecName: "kube-api-access-bg99v") pod "2fe229a6-e9e2-40b3-a4d7-0780c4f5c884" (UID: "2fe229a6-e9e2-40b3-a4d7-0780c4f5c884"). InnerVolumeSpecName "kube-api-access-bg99v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.689882 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "2fe229a6-e9e2-40b3-a4d7-0780c4f5c884" (UID: "2fe229a6-e9e2-40b3-a4d7-0780c4f5c884"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.783908 4809 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.783947 4809 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.783959 4809 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.783987 4809 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.783999 4809 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.784010 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:41 crc kubenswrapper[4809]: I1124 07:06:41.784021 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bg99v\" (UniqueName: \"kubernetes.io/projected/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884-kube-api-access-bg99v\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:42 crc kubenswrapper[4809]: I1124 07:06:42.266119 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-8r5vw_2fe229a6-e9e2-40b3-a4d7-0780c4f5c884/console/0.log" Nov 24 07:06:42 crc kubenswrapper[4809]: I1124 07:06:42.266169 4809 generic.go:334] "Generic (PLEG): container finished" podID="2fe229a6-e9e2-40b3-a4d7-0780c4f5c884" containerID="a167cf5d10e05e3dc032dd5a98a000f04efc29c0ee74ed9e382c4405089d041f" exitCode=2 Nov 24 07:06:42 crc kubenswrapper[4809]: I1124 07:06:42.266347 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8r5vw" Nov 24 07:06:42 crc kubenswrapper[4809]: I1124 07:06:42.267010 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8r5vw" event={"ID":"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884","Type":"ContainerDied","Data":"a167cf5d10e05e3dc032dd5a98a000f04efc29c0ee74ed9e382c4405089d041f"} Nov 24 07:06:42 crc kubenswrapper[4809]: I1124 07:06:42.267050 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8r5vw" event={"ID":"2fe229a6-e9e2-40b3-a4d7-0780c4f5c884","Type":"ContainerDied","Data":"55a4bf0e862c2ea68500c04a7e69e8261fc4d00c00343076a1cef7ecaaa25189"} Nov 24 07:06:42 crc kubenswrapper[4809]: I1124 07:06:42.267068 4809 scope.go:117] "RemoveContainer" containerID="a167cf5d10e05e3dc032dd5a98a000f04efc29c0ee74ed9e382c4405089d041f" Nov 24 07:06:42 crc kubenswrapper[4809]: I1124 07:06:42.292095 4809 scope.go:117] "RemoveContainer" containerID="a167cf5d10e05e3dc032dd5a98a000f04efc29c0ee74ed9e382c4405089d041f" Nov 24 07:06:42 crc kubenswrapper[4809]: E1124 07:06:42.293501 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a167cf5d10e05e3dc032dd5a98a000f04efc29c0ee74ed9e382c4405089d041f\": container with ID starting with a167cf5d10e05e3dc032dd5a98a000f04efc29c0ee74ed9e382c4405089d041f not found: ID does not exist" containerID="a167cf5d10e05e3dc032dd5a98a000f04efc29c0ee74ed9e382c4405089d041f" Nov 24 07:06:42 crc kubenswrapper[4809]: I1124 07:06:42.293587 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a167cf5d10e05e3dc032dd5a98a000f04efc29c0ee74ed9e382c4405089d041f"} err="failed to get container status \"a167cf5d10e05e3dc032dd5a98a000f04efc29c0ee74ed9e382c4405089d041f\": rpc error: code = NotFound desc = could not find container \"a167cf5d10e05e3dc032dd5a98a000f04efc29c0ee74ed9e382c4405089d041f\": container with ID starting with a167cf5d10e05e3dc032dd5a98a000f04efc29c0ee74ed9e382c4405089d041f not found: ID does not exist" Nov 24 07:06:42 crc kubenswrapper[4809]: I1124 07:06:42.297758 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-8r5vw"] Nov 24 07:06:42 crc kubenswrapper[4809]: I1124 07:06:42.300189 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-8r5vw"] Nov 24 07:06:42 crc kubenswrapper[4809]: I1124 07:06:42.470987 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8" Nov 24 07:06:42 crc kubenswrapper[4809]: I1124 07:06:42.492878 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw7mh\" (UniqueName: \"kubernetes.io/projected/7feb42ed-e415-44b8-a837-da74daf4aa52-kube-api-access-bw7mh\") pod \"7feb42ed-e415-44b8-a837-da74daf4aa52\" (UID: \"7feb42ed-e415-44b8-a837-da74daf4aa52\") " Nov 24 07:06:42 crc kubenswrapper[4809]: I1124 07:06:42.492934 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7feb42ed-e415-44b8-a837-da74daf4aa52-util\") pod \"7feb42ed-e415-44b8-a837-da74daf4aa52\" (UID: \"7feb42ed-e415-44b8-a837-da74daf4aa52\") " Nov 24 07:06:42 crc kubenswrapper[4809]: I1124 07:06:42.498552 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7feb42ed-e415-44b8-a837-da74daf4aa52-kube-api-access-bw7mh" (OuterVolumeSpecName: "kube-api-access-bw7mh") pod "7feb42ed-e415-44b8-a837-da74daf4aa52" (UID: "7feb42ed-e415-44b8-a837-da74daf4aa52"). InnerVolumeSpecName "kube-api-access-bw7mh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:06:42 crc kubenswrapper[4809]: I1124 07:06:42.506753 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7feb42ed-e415-44b8-a837-da74daf4aa52-util" (OuterVolumeSpecName: "util") pod "7feb42ed-e415-44b8-a837-da74daf4aa52" (UID: "7feb42ed-e415-44b8-a837-da74daf4aa52"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:06:42 crc kubenswrapper[4809]: I1124 07:06:42.594283 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7feb42ed-e415-44b8-a837-da74daf4aa52-bundle\") pod \"7feb42ed-e415-44b8-a837-da74daf4aa52\" (UID: \"7feb42ed-e415-44b8-a837-da74daf4aa52\") " Nov 24 07:06:42 crc kubenswrapper[4809]: I1124 07:06:42.594560 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw7mh\" (UniqueName: \"kubernetes.io/projected/7feb42ed-e415-44b8-a837-da74daf4aa52-kube-api-access-bw7mh\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:42 crc kubenswrapper[4809]: I1124 07:06:42.594573 4809 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7feb42ed-e415-44b8-a837-da74daf4aa52-util\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:42 crc kubenswrapper[4809]: I1124 07:06:42.595644 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7feb42ed-e415-44b8-a837-da74daf4aa52-bundle" (OuterVolumeSpecName: "bundle") pod "7feb42ed-e415-44b8-a837-da74daf4aa52" (UID: "7feb42ed-e415-44b8-a837-da74daf4aa52"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:06:42 crc kubenswrapper[4809]: I1124 07:06:42.696789 4809 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7feb42ed-e415-44b8-a837-da74daf4aa52-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:42 crc kubenswrapper[4809]: I1124 07:06:42.898171 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fe229a6-e9e2-40b3-a4d7-0780c4f5c884" path="/var/lib/kubelet/pods/2fe229a6-e9e2-40b3-a4d7-0780c4f5c884/volumes" Nov 24 07:06:43 crc kubenswrapper[4809]: I1124 07:06:43.273776 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8" Nov 24 07:06:43 crc kubenswrapper[4809]: I1124 07:06:43.273757 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8" event={"ID":"7feb42ed-e415-44b8-a837-da74daf4aa52","Type":"ContainerDied","Data":"b78533784f968363c5b7dce1f568865e6c8a44e54a8d07b64838f61514f80a1f"} Nov 24 07:06:43 crc kubenswrapper[4809]: I1124 07:06:43.274223 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b78533784f968363c5b7dce1f568865e6c8a44e54a8d07b64838f61514f80a1f" Nov 24 07:06:48 crc kubenswrapper[4809]: I1124 07:06:48.043501 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:06:48 crc kubenswrapper[4809]: I1124 07:06:48.043800 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.167133 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-7d8556447d-spn4w"] Nov 24 07:06:53 crc kubenswrapper[4809]: E1124 07:06:53.167607 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7feb42ed-e415-44b8-a837-da74daf4aa52" containerName="extract" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.167620 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7feb42ed-e415-44b8-a837-da74daf4aa52" containerName="extract" Nov 24 07:06:53 crc kubenswrapper[4809]: E1124 07:06:53.167630 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7feb42ed-e415-44b8-a837-da74daf4aa52" containerName="util" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.167636 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7feb42ed-e415-44b8-a837-da74daf4aa52" containerName="util" Nov 24 07:06:53 crc kubenswrapper[4809]: E1124 07:06:53.167645 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7feb42ed-e415-44b8-a837-da74daf4aa52" containerName="pull" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.167650 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7feb42ed-e415-44b8-a837-da74daf4aa52" containerName="pull" Nov 24 07:06:53 crc kubenswrapper[4809]: E1124 07:06:53.167663 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fe229a6-e9e2-40b3-a4d7-0780c4f5c884" containerName="console" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.167669 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fe229a6-e9e2-40b3-a4d7-0780c4f5c884" containerName="console" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.167755 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7feb42ed-e415-44b8-a837-da74daf4aa52" containerName="extract" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.167767 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fe229a6-e9e2-40b3-a4d7-0780c4f5c884" containerName="console" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.168157 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7d8556447d-spn4w" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.169444 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.178274 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.179757 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.180146 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-4dbdx" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.180212 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.186666 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7d8556447d-spn4w"] Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.212985 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a9a36b6c-e79c-4180-9118-146f3d280f15-webhook-cert\") pod \"metallb-operator-controller-manager-7d8556447d-spn4w\" (UID: \"a9a36b6c-e79c-4180-9118-146f3d280f15\") " pod="metallb-system/metallb-operator-controller-manager-7d8556447d-spn4w" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.213037 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a9a36b6c-e79c-4180-9118-146f3d280f15-apiservice-cert\") pod \"metallb-operator-controller-manager-7d8556447d-spn4w\" (UID: \"a9a36b6c-e79c-4180-9118-146f3d280f15\") " pod="metallb-system/metallb-operator-controller-manager-7d8556447d-spn4w" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.213080 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qqtt\" (UniqueName: \"kubernetes.io/projected/a9a36b6c-e79c-4180-9118-146f3d280f15-kube-api-access-2qqtt\") pod \"metallb-operator-controller-manager-7d8556447d-spn4w\" (UID: \"a9a36b6c-e79c-4180-9118-146f3d280f15\") " pod="metallb-system/metallb-operator-controller-manager-7d8556447d-spn4w" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.313728 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a9a36b6c-e79c-4180-9118-146f3d280f15-apiservice-cert\") pod \"metallb-operator-controller-manager-7d8556447d-spn4w\" (UID: \"a9a36b6c-e79c-4180-9118-146f3d280f15\") " pod="metallb-system/metallb-operator-controller-manager-7d8556447d-spn4w" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.313790 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qqtt\" (UniqueName: \"kubernetes.io/projected/a9a36b6c-e79c-4180-9118-146f3d280f15-kube-api-access-2qqtt\") pod \"metallb-operator-controller-manager-7d8556447d-spn4w\" (UID: \"a9a36b6c-e79c-4180-9118-146f3d280f15\") " pod="metallb-system/metallb-operator-controller-manager-7d8556447d-spn4w" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.313843 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a9a36b6c-e79c-4180-9118-146f3d280f15-webhook-cert\") pod \"metallb-operator-controller-manager-7d8556447d-spn4w\" (UID: \"a9a36b6c-e79c-4180-9118-146f3d280f15\") " pod="metallb-system/metallb-operator-controller-manager-7d8556447d-spn4w" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.323778 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a9a36b6c-e79c-4180-9118-146f3d280f15-apiservice-cert\") pod \"metallb-operator-controller-manager-7d8556447d-spn4w\" (UID: \"a9a36b6c-e79c-4180-9118-146f3d280f15\") " pod="metallb-system/metallb-operator-controller-manager-7d8556447d-spn4w" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.323802 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a9a36b6c-e79c-4180-9118-146f3d280f15-webhook-cert\") pod \"metallb-operator-controller-manager-7d8556447d-spn4w\" (UID: \"a9a36b6c-e79c-4180-9118-146f3d280f15\") " pod="metallb-system/metallb-operator-controller-manager-7d8556447d-spn4w" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.331544 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qqtt\" (UniqueName: \"kubernetes.io/projected/a9a36b6c-e79c-4180-9118-146f3d280f15-kube-api-access-2qqtt\") pod \"metallb-operator-controller-manager-7d8556447d-spn4w\" (UID: \"a9a36b6c-e79c-4180-9118-146f3d280f15\") " pod="metallb-system/metallb-operator-controller-manager-7d8556447d-spn4w" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.400813 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5969b4cdf8-zmvtt"] Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.401454 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5969b4cdf8-zmvtt" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.408151 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.407953 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.410175 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-8pg74" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.413887 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5969b4cdf8-zmvtt"] Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.414778 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/baa671e1-bc87-45e6-8c04-33f408289fb4-webhook-cert\") pod \"metallb-operator-webhook-server-5969b4cdf8-zmvtt\" (UID: \"baa671e1-bc87-45e6-8c04-33f408289fb4\") " pod="metallb-system/metallb-operator-webhook-server-5969b4cdf8-zmvtt" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.414921 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/baa671e1-bc87-45e6-8c04-33f408289fb4-apiservice-cert\") pod \"metallb-operator-webhook-server-5969b4cdf8-zmvtt\" (UID: \"baa671e1-bc87-45e6-8c04-33f408289fb4\") " pod="metallb-system/metallb-operator-webhook-server-5969b4cdf8-zmvtt" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.415024 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v59pv\" (UniqueName: \"kubernetes.io/projected/baa671e1-bc87-45e6-8c04-33f408289fb4-kube-api-access-v59pv\") pod \"metallb-operator-webhook-server-5969b4cdf8-zmvtt\" (UID: \"baa671e1-bc87-45e6-8c04-33f408289fb4\") " pod="metallb-system/metallb-operator-webhook-server-5969b4cdf8-zmvtt" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.481768 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7d8556447d-spn4w" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.515655 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/baa671e1-bc87-45e6-8c04-33f408289fb4-webhook-cert\") pod \"metallb-operator-webhook-server-5969b4cdf8-zmvtt\" (UID: \"baa671e1-bc87-45e6-8c04-33f408289fb4\") " pod="metallb-system/metallb-operator-webhook-server-5969b4cdf8-zmvtt" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.515946 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/baa671e1-bc87-45e6-8c04-33f408289fb4-apiservice-cert\") pod \"metallb-operator-webhook-server-5969b4cdf8-zmvtt\" (UID: \"baa671e1-bc87-45e6-8c04-33f408289fb4\") " pod="metallb-system/metallb-operator-webhook-server-5969b4cdf8-zmvtt" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.516079 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v59pv\" (UniqueName: \"kubernetes.io/projected/baa671e1-bc87-45e6-8c04-33f408289fb4-kube-api-access-v59pv\") pod \"metallb-operator-webhook-server-5969b4cdf8-zmvtt\" (UID: \"baa671e1-bc87-45e6-8c04-33f408289fb4\") " pod="metallb-system/metallb-operator-webhook-server-5969b4cdf8-zmvtt" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.525219 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/baa671e1-bc87-45e6-8c04-33f408289fb4-apiservice-cert\") pod \"metallb-operator-webhook-server-5969b4cdf8-zmvtt\" (UID: \"baa671e1-bc87-45e6-8c04-33f408289fb4\") " pod="metallb-system/metallb-operator-webhook-server-5969b4cdf8-zmvtt" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.533736 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/baa671e1-bc87-45e6-8c04-33f408289fb4-webhook-cert\") pod \"metallb-operator-webhook-server-5969b4cdf8-zmvtt\" (UID: \"baa671e1-bc87-45e6-8c04-33f408289fb4\") " pod="metallb-system/metallb-operator-webhook-server-5969b4cdf8-zmvtt" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.536530 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v59pv\" (UniqueName: \"kubernetes.io/projected/baa671e1-bc87-45e6-8c04-33f408289fb4-kube-api-access-v59pv\") pod \"metallb-operator-webhook-server-5969b4cdf8-zmvtt\" (UID: \"baa671e1-bc87-45e6-8c04-33f408289fb4\") " pod="metallb-system/metallb-operator-webhook-server-5969b4cdf8-zmvtt" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.715938 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5969b4cdf8-zmvtt" Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.915608 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5969b4cdf8-zmvtt"] Nov 24 07:06:53 crc kubenswrapper[4809]: I1124 07:06:53.920810 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7d8556447d-spn4w"] Nov 24 07:06:53 crc kubenswrapper[4809]: W1124 07:06:53.922723 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9a36b6c_e79c_4180_9118_146f3d280f15.slice/crio-34859e1bf6e0335d16f817c485786dec1b2d5f5a433733ba6adb546030627132 WatchSource:0}: Error finding container 34859e1bf6e0335d16f817c485786dec1b2d5f5a433733ba6adb546030627132: Status 404 returned error can't find the container with id 34859e1bf6e0335d16f817c485786dec1b2d5f5a433733ba6adb546030627132 Nov 24 07:06:54 crc kubenswrapper[4809]: I1124 07:06:54.340878 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7d8556447d-spn4w" event={"ID":"a9a36b6c-e79c-4180-9118-146f3d280f15","Type":"ContainerStarted","Data":"34859e1bf6e0335d16f817c485786dec1b2d5f5a433733ba6adb546030627132"} Nov 24 07:06:54 crc kubenswrapper[4809]: I1124 07:06:54.342103 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5969b4cdf8-zmvtt" event={"ID":"baa671e1-bc87-45e6-8c04-33f408289fb4","Type":"ContainerStarted","Data":"4e66be6383c46489ecaec52d78ca7bf530687d190b95fcbc6229c15d529e0c83"} Nov 24 07:06:58 crc kubenswrapper[4809]: I1124 07:06:58.372950 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5969b4cdf8-zmvtt" event={"ID":"baa671e1-bc87-45e6-8c04-33f408289fb4","Type":"ContainerStarted","Data":"83068445401bd73281d225ca77ba9ed2a0d1cc640599f2f92a9cfc6366194b17"} Nov 24 07:06:58 crc kubenswrapper[4809]: I1124 07:06:58.373460 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5969b4cdf8-zmvtt" Nov 24 07:06:58 crc kubenswrapper[4809]: I1124 07:06:58.375694 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7d8556447d-spn4w" event={"ID":"a9a36b6c-e79c-4180-9118-146f3d280f15","Type":"ContainerStarted","Data":"833c6a49794655d44cd9be4a00cc46bacac937c5ac4d513ae1ba457483758f11"} Nov 24 07:06:58 crc kubenswrapper[4809]: I1124 07:06:58.376017 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-7d8556447d-spn4w" Nov 24 07:06:58 crc kubenswrapper[4809]: I1124 07:06:58.396903 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5969b4cdf8-zmvtt" podStartSLOduration=1.231143816 podStartE2EDuration="5.396883051s" podCreationTimestamp="2025-11-24 07:06:53 +0000 UTC" firstStartedPulling="2025-11-24 07:06:53.922143022 +0000 UTC m=+713.822734727" lastFinishedPulling="2025-11-24 07:06:58.087882257 +0000 UTC m=+717.988473962" observedRunningTime="2025-11-24 07:06:58.395600615 +0000 UTC m=+718.296192330" watchObservedRunningTime="2025-11-24 07:06:58.396883051 +0000 UTC m=+718.297474766" Nov 24 07:06:58 crc kubenswrapper[4809]: I1124 07:06:58.421309 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-7d8556447d-spn4w" podStartSLOduration=1.274427049 podStartE2EDuration="5.421282932s" podCreationTimestamp="2025-11-24 07:06:53 +0000 UTC" firstStartedPulling="2025-11-24 07:06:53.924615229 +0000 UTC m=+713.825206934" lastFinishedPulling="2025-11-24 07:06:58.071471112 +0000 UTC m=+717.972062817" observedRunningTime="2025-11-24 07:06:58.415835714 +0000 UTC m=+718.316427439" watchObservedRunningTime="2025-11-24 07:06:58.421282932 +0000 UTC m=+718.321874657" Nov 24 07:07:13 crc kubenswrapper[4809]: I1124 07:07:13.725116 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5969b4cdf8-zmvtt" Nov 24 07:07:18 crc kubenswrapper[4809]: I1124 07:07:18.043551 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:07:18 crc kubenswrapper[4809]: I1124 07:07:18.043610 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.409360 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-92k7t"] Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.410269 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" podUID="f3ac0673-5554-4ac6-a16e-b08631fdbf94" containerName="controller-manager" containerID="cri-o://e33de40b31ff2c23d7d9a91492598171a4d75feb1523db57b411336a7596464f" gracePeriod=30 Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.513945 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl"] Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.514221 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" podUID="358682a8-6b18-49ee-bd7d-d9eab2b22ee0" containerName="route-controller-manager" containerID="cri-o://fc4e9365dbecfca96183bb8ac808a45f2fae8044e741583d2c8e1bcd913f9d74" gracePeriod=30 Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.820839 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.868894 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.948480 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4r4sq\" (UniqueName: \"kubernetes.io/projected/f3ac0673-5554-4ac6-a16e-b08631fdbf94-kube-api-access-4r4sq\") pod \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\" (UID: \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\") " Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.949237 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3ac0673-5554-4ac6-a16e-b08631fdbf94-serving-cert\") pod \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\" (UID: \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\") " Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.949324 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f3ac0673-5554-4ac6-a16e-b08631fdbf94-client-ca\") pod \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\" (UID: \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\") " Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.949410 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wr6gp\" (UniqueName: \"kubernetes.io/projected/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-kube-api-access-wr6gp\") pod \"358682a8-6b18-49ee-bd7d-d9eab2b22ee0\" (UID: \"358682a8-6b18-49ee-bd7d-d9eab2b22ee0\") " Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.949461 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-serving-cert\") pod \"358682a8-6b18-49ee-bd7d-d9eab2b22ee0\" (UID: \"358682a8-6b18-49ee-bd7d-d9eab2b22ee0\") " Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.949529 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3ac0673-5554-4ac6-a16e-b08631fdbf94-config\") pod \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\" (UID: \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\") " Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.949579 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f3ac0673-5554-4ac6-a16e-b08631fdbf94-proxy-ca-bundles\") pod \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\" (UID: \"f3ac0673-5554-4ac6-a16e-b08631fdbf94\") " Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.949621 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-config\") pod \"358682a8-6b18-49ee-bd7d-d9eab2b22ee0\" (UID: \"358682a8-6b18-49ee-bd7d-d9eab2b22ee0\") " Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.949656 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-client-ca\") pod \"358682a8-6b18-49ee-bd7d-d9eab2b22ee0\" (UID: \"358682a8-6b18-49ee-bd7d-d9eab2b22ee0\") " Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.950023 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3ac0673-5554-4ac6-a16e-b08631fdbf94-client-ca" (OuterVolumeSpecName: "client-ca") pod "f3ac0673-5554-4ac6-a16e-b08631fdbf94" (UID: "f3ac0673-5554-4ac6-a16e-b08631fdbf94"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.950165 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3ac0673-5554-4ac6-a16e-b08631fdbf94-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "f3ac0673-5554-4ac6-a16e-b08631fdbf94" (UID: "f3ac0673-5554-4ac6-a16e-b08631fdbf94"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.950269 4809 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f3ac0673-5554-4ac6-a16e-b08631fdbf94-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.950814 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3ac0673-5554-4ac6-a16e-b08631fdbf94-config" (OuterVolumeSpecName: "config") pod "f3ac0673-5554-4ac6-a16e-b08631fdbf94" (UID: "f3ac0673-5554-4ac6-a16e-b08631fdbf94"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.950837 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-client-ca" (OuterVolumeSpecName: "client-ca") pod "358682a8-6b18-49ee-bd7d-d9eab2b22ee0" (UID: "358682a8-6b18-49ee-bd7d-d9eab2b22ee0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.950898 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-config" (OuterVolumeSpecName: "config") pod "358682a8-6b18-49ee-bd7d-d9eab2b22ee0" (UID: "358682a8-6b18-49ee-bd7d-d9eab2b22ee0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.955341 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "358682a8-6b18-49ee-bd7d-d9eab2b22ee0" (UID: "358682a8-6b18-49ee-bd7d-d9eab2b22ee0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.955374 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3ac0673-5554-4ac6-a16e-b08631fdbf94-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f3ac0673-5554-4ac6-a16e-b08631fdbf94" (UID: "f3ac0673-5554-4ac6-a16e-b08631fdbf94"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.955374 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-kube-api-access-wr6gp" (OuterVolumeSpecName: "kube-api-access-wr6gp") pod "358682a8-6b18-49ee-bd7d-d9eab2b22ee0" (UID: "358682a8-6b18-49ee-bd7d-d9eab2b22ee0"). InnerVolumeSpecName "kube-api-access-wr6gp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:07:26 crc kubenswrapper[4809]: I1124 07:07:26.955896 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3ac0673-5554-4ac6-a16e-b08631fdbf94-kube-api-access-4r4sq" (OuterVolumeSpecName: "kube-api-access-4r4sq") pod "f3ac0673-5554-4ac6-a16e-b08631fdbf94" (UID: "f3ac0673-5554-4ac6-a16e-b08631fdbf94"). InnerVolumeSpecName "kube-api-access-4r4sq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.050959 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wr6gp\" (UniqueName: \"kubernetes.io/projected/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-kube-api-access-wr6gp\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.051025 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.051065 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3ac0673-5554-4ac6-a16e-b08631fdbf94-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.051079 4809 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f3ac0673-5554-4ac6-a16e-b08631fdbf94-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.051091 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.051101 4809 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/358682a8-6b18-49ee-bd7d-d9eab2b22ee0-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.051113 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4r4sq\" (UniqueName: \"kubernetes.io/projected/f3ac0673-5554-4ac6-a16e-b08631fdbf94-kube-api-access-4r4sq\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.051124 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3ac0673-5554-4ac6-a16e-b08631fdbf94-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.527153 4809 generic.go:334] "Generic (PLEG): container finished" podID="358682a8-6b18-49ee-bd7d-d9eab2b22ee0" containerID="fc4e9365dbecfca96183bb8ac808a45f2fae8044e741583d2c8e1bcd913f9d74" exitCode=0 Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.527215 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" event={"ID":"358682a8-6b18-49ee-bd7d-d9eab2b22ee0","Type":"ContainerDied","Data":"fc4e9365dbecfca96183bb8ac808a45f2fae8044e741583d2c8e1bcd913f9d74"} Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.527242 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" event={"ID":"358682a8-6b18-49ee-bd7d-d9eab2b22ee0","Type":"ContainerDied","Data":"713f6b38bee99b89dc7907e191f636f646dda0aa80d016e0fc10e719f3843c75"} Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.527257 4809 scope.go:117] "RemoveContainer" containerID="fc4e9365dbecfca96183bb8ac808a45f2fae8044e741583d2c8e1bcd913f9d74" Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.527353 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl" Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.533829 4809 generic.go:334] "Generic (PLEG): container finished" podID="f3ac0673-5554-4ac6-a16e-b08631fdbf94" containerID="e33de40b31ff2c23d7d9a91492598171a4d75feb1523db57b411336a7596464f" exitCode=0 Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.534220 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" event={"ID":"f3ac0673-5554-4ac6-a16e-b08631fdbf94","Type":"ContainerDied","Data":"e33de40b31ff2c23d7d9a91492598171a4d75feb1523db57b411336a7596464f"} Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.534259 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" event={"ID":"f3ac0673-5554-4ac6-a16e-b08631fdbf94","Type":"ContainerDied","Data":"cc23e664312807f16fb1c5601b42d4544c2e8d9fb7a5acf7afc28219cfafbe3d"} Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.534348 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-92k7t" Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.577851 4809 scope.go:117] "RemoveContainer" containerID="fc4e9365dbecfca96183bb8ac808a45f2fae8044e741583d2c8e1bcd913f9d74" Nov 24 07:07:27 crc kubenswrapper[4809]: E1124 07:07:27.578331 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc4e9365dbecfca96183bb8ac808a45f2fae8044e741583d2c8e1bcd913f9d74\": container with ID starting with fc4e9365dbecfca96183bb8ac808a45f2fae8044e741583d2c8e1bcd913f9d74 not found: ID does not exist" containerID="fc4e9365dbecfca96183bb8ac808a45f2fae8044e741583d2c8e1bcd913f9d74" Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.578371 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc4e9365dbecfca96183bb8ac808a45f2fae8044e741583d2c8e1bcd913f9d74"} err="failed to get container status \"fc4e9365dbecfca96183bb8ac808a45f2fae8044e741583d2c8e1bcd913f9d74\": rpc error: code = NotFound desc = could not find container \"fc4e9365dbecfca96183bb8ac808a45f2fae8044e741583d2c8e1bcd913f9d74\": container with ID starting with fc4e9365dbecfca96183bb8ac808a45f2fae8044e741583d2c8e1bcd913f9d74 not found: ID does not exist" Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.578396 4809 scope.go:117] "RemoveContainer" containerID="e33de40b31ff2c23d7d9a91492598171a4d75feb1523db57b411336a7596464f" Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.585156 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl"] Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.590367 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-p86rl"] Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.591717 4809 scope.go:117] "RemoveContainer" containerID="e33de40b31ff2c23d7d9a91492598171a4d75feb1523db57b411336a7596464f" Nov 24 07:07:27 crc kubenswrapper[4809]: E1124 07:07:27.592088 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e33de40b31ff2c23d7d9a91492598171a4d75feb1523db57b411336a7596464f\": container with ID starting with e33de40b31ff2c23d7d9a91492598171a4d75feb1523db57b411336a7596464f not found: ID does not exist" containerID="e33de40b31ff2c23d7d9a91492598171a4d75feb1523db57b411336a7596464f" Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.592114 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e33de40b31ff2c23d7d9a91492598171a4d75feb1523db57b411336a7596464f"} err="failed to get container status \"e33de40b31ff2c23d7d9a91492598171a4d75feb1523db57b411336a7596464f\": rpc error: code = NotFound desc = could not find container \"e33de40b31ff2c23d7d9a91492598171a4d75feb1523db57b411336a7596464f\": container with ID starting with e33de40b31ff2c23d7d9a91492598171a4d75feb1523db57b411336a7596464f not found: ID does not exist" Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.599317 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-92k7t"] Nov 24 07:07:27 crc kubenswrapper[4809]: I1124 07:07:27.601993 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-92k7t"] Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.426120 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-84d4cc4998-msdhj"] Nov 24 07:07:28 crc kubenswrapper[4809]: E1124 07:07:28.426478 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3ac0673-5554-4ac6-a16e-b08631fdbf94" containerName="controller-manager" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.426503 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3ac0673-5554-4ac6-a16e-b08631fdbf94" containerName="controller-manager" Nov 24 07:07:28 crc kubenswrapper[4809]: E1124 07:07:28.426529 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="358682a8-6b18-49ee-bd7d-d9eab2b22ee0" containerName="route-controller-manager" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.426541 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="358682a8-6b18-49ee-bd7d-d9eab2b22ee0" containerName="route-controller-manager" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.426666 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="358682a8-6b18-49ee-bd7d-d9eab2b22ee0" containerName="route-controller-manager" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.426685 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3ac0673-5554-4ac6-a16e-b08631fdbf94" containerName="controller-manager" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.427256 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-84d4cc4998-msdhj" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.430234 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.430301 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fbf59846b-zc9sg"] Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.431137 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7fbf59846b-zc9sg" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.432325 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.432645 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.433026 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.433140 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.434996 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.435086 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.435404 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.435438 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.436038 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.436059 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.436325 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.446731 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.450436 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fbf59846b-zc9sg"] Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.483163 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-84d4cc4998-msdhj"] Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.572714 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a48066b-1019-4c99-9c1e-09958fba15cb-serving-cert\") pod \"route-controller-manager-7fbf59846b-zc9sg\" (UID: \"4a48066b-1019-4c99-9c1e-09958fba15cb\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf59846b-zc9sg" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.572772 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4a48066b-1019-4c99-9c1e-09958fba15cb-client-ca\") pod \"route-controller-manager-7fbf59846b-zc9sg\" (UID: \"4a48066b-1019-4c99-9c1e-09958fba15cb\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf59846b-zc9sg" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.572801 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b7a5653-eaf1-409a-a343-1e1ffb898b6b-serving-cert\") pod \"controller-manager-84d4cc4998-msdhj\" (UID: \"3b7a5653-eaf1-409a-a343-1e1ffb898b6b\") " pod="openshift-controller-manager/controller-manager-84d4cc4998-msdhj" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.572821 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a48066b-1019-4c99-9c1e-09958fba15cb-config\") pod \"route-controller-manager-7fbf59846b-zc9sg\" (UID: \"4a48066b-1019-4c99-9c1e-09958fba15cb\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf59846b-zc9sg" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.572844 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3b7a5653-eaf1-409a-a343-1e1ffb898b6b-client-ca\") pod \"controller-manager-84d4cc4998-msdhj\" (UID: \"3b7a5653-eaf1-409a-a343-1e1ffb898b6b\") " pod="openshift-controller-manager/controller-manager-84d4cc4998-msdhj" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.572864 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw9vn\" (UniqueName: \"kubernetes.io/projected/4a48066b-1019-4c99-9c1e-09958fba15cb-kube-api-access-rw9vn\") pod \"route-controller-manager-7fbf59846b-zc9sg\" (UID: \"4a48066b-1019-4c99-9c1e-09958fba15cb\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf59846b-zc9sg" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.572887 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3b7a5653-eaf1-409a-a343-1e1ffb898b6b-proxy-ca-bundles\") pod \"controller-manager-84d4cc4998-msdhj\" (UID: \"3b7a5653-eaf1-409a-a343-1e1ffb898b6b\") " pod="openshift-controller-manager/controller-manager-84d4cc4998-msdhj" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.572917 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b7a5653-eaf1-409a-a343-1e1ffb898b6b-config\") pod \"controller-manager-84d4cc4998-msdhj\" (UID: \"3b7a5653-eaf1-409a-a343-1e1ffb898b6b\") " pod="openshift-controller-manager/controller-manager-84d4cc4998-msdhj" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.572933 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjbv2\" (UniqueName: \"kubernetes.io/projected/3b7a5653-eaf1-409a-a343-1e1ffb898b6b-kube-api-access-fjbv2\") pod \"controller-manager-84d4cc4998-msdhj\" (UID: \"3b7a5653-eaf1-409a-a343-1e1ffb898b6b\") " pod="openshift-controller-manager/controller-manager-84d4cc4998-msdhj" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.674292 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b7a5653-eaf1-409a-a343-1e1ffb898b6b-config\") pod \"controller-manager-84d4cc4998-msdhj\" (UID: \"3b7a5653-eaf1-409a-a343-1e1ffb898b6b\") " pod="openshift-controller-manager/controller-manager-84d4cc4998-msdhj" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.674337 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjbv2\" (UniqueName: \"kubernetes.io/projected/3b7a5653-eaf1-409a-a343-1e1ffb898b6b-kube-api-access-fjbv2\") pod \"controller-manager-84d4cc4998-msdhj\" (UID: \"3b7a5653-eaf1-409a-a343-1e1ffb898b6b\") " pod="openshift-controller-manager/controller-manager-84d4cc4998-msdhj" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.674377 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a48066b-1019-4c99-9c1e-09958fba15cb-serving-cert\") pod \"route-controller-manager-7fbf59846b-zc9sg\" (UID: \"4a48066b-1019-4c99-9c1e-09958fba15cb\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf59846b-zc9sg" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.674441 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4a48066b-1019-4c99-9c1e-09958fba15cb-client-ca\") pod \"route-controller-manager-7fbf59846b-zc9sg\" (UID: \"4a48066b-1019-4c99-9c1e-09958fba15cb\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf59846b-zc9sg" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.674491 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b7a5653-eaf1-409a-a343-1e1ffb898b6b-serving-cert\") pod \"controller-manager-84d4cc4998-msdhj\" (UID: \"3b7a5653-eaf1-409a-a343-1e1ffb898b6b\") " pod="openshift-controller-manager/controller-manager-84d4cc4998-msdhj" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.674512 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a48066b-1019-4c99-9c1e-09958fba15cb-config\") pod \"route-controller-manager-7fbf59846b-zc9sg\" (UID: \"4a48066b-1019-4c99-9c1e-09958fba15cb\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf59846b-zc9sg" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.674551 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3b7a5653-eaf1-409a-a343-1e1ffb898b6b-client-ca\") pod \"controller-manager-84d4cc4998-msdhj\" (UID: \"3b7a5653-eaf1-409a-a343-1e1ffb898b6b\") " pod="openshift-controller-manager/controller-manager-84d4cc4998-msdhj" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.674577 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw9vn\" (UniqueName: \"kubernetes.io/projected/4a48066b-1019-4c99-9c1e-09958fba15cb-kube-api-access-rw9vn\") pod \"route-controller-manager-7fbf59846b-zc9sg\" (UID: \"4a48066b-1019-4c99-9c1e-09958fba15cb\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf59846b-zc9sg" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.674609 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3b7a5653-eaf1-409a-a343-1e1ffb898b6b-proxy-ca-bundles\") pod \"controller-manager-84d4cc4998-msdhj\" (UID: \"3b7a5653-eaf1-409a-a343-1e1ffb898b6b\") " pod="openshift-controller-manager/controller-manager-84d4cc4998-msdhj" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.675507 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3b7a5653-eaf1-409a-a343-1e1ffb898b6b-proxy-ca-bundles\") pod \"controller-manager-84d4cc4998-msdhj\" (UID: \"3b7a5653-eaf1-409a-a343-1e1ffb898b6b\") " pod="openshift-controller-manager/controller-manager-84d4cc4998-msdhj" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.675957 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4a48066b-1019-4c99-9c1e-09958fba15cb-client-ca\") pod \"route-controller-manager-7fbf59846b-zc9sg\" (UID: \"4a48066b-1019-4c99-9c1e-09958fba15cb\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf59846b-zc9sg" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.676099 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b7a5653-eaf1-409a-a343-1e1ffb898b6b-config\") pod \"controller-manager-84d4cc4998-msdhj\" (UID: \"3b7a5653-eaf1-409a-a343-1e1ffb898b6b\") " pod="openshift-controller-manager/controller-manager-84d4cc4998-msdhj" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.676424 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3b7a5653-eaf1-409a-a343-1e1ffb898b6b-client-ca\") pod \"controller-manager-84d4cc4998-msdhj\" (UID: \"3b7a5653-eaf1-409a-a343-1e1ffb898b6b\") " pod="openshift-controller-manager/controller-manager-84d4cc4998-msdhj" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.676588 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a48066b-1019-4c99-9c1e-09958fba15cb-config\") pod \"route-controller-manager-7fbf59846b-zc9sg\" (UID: \"4a48066b-1019-4c99-9c1e-09958fba15cb\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf59846b-zc9sg" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.680351 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a48066b-1019-4c99-9c1e-09958fba15cb-serving-cert\") pod \"route-controller-manager-7fbf59846b-zc9sg\" (UID: \"4a48066b-1019-4c99-9c1e-09958fba15cb\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf59846b-zc9sg" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.680739 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b7a5653-eaf1-409a-a343-1e1ffb898b6b-serving-cert\") pod \"controller-manager-84d4cc4998-msdhj\" (UID: \"3b7a5653-eaf1-409a-a343-1e1ffb898b6b\") " pod="openshift-controller-manager/controller-manager-84d4cc4998-msdhj" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.692421 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjbv2\" (UniqueName: \"kubernetes.io/projected/3b7a5653-eaf1-409a-a343-1e1ffb898b6b-kube-api-access-fjbv2\") pod \"controller-manager-84d4cc4998-msdhj\" (UID: \"3b7a5653-eaf1-409a-a343-1e1ffb898b6b\") " pod="openshift-controller-manager/controller-manager-84d4cc4998-msdhj" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.694914 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw9vn\" (UniqueName: \"kubernetes.io/projected/4a48066b-1019-4c99-9c1e-09958fba15cb-kube-api-access-rw9vn\") pod \"route-controller-manager-7fbf59846b-zc9sg\" (UID: \"4a48066b-1019-4c99-9c1e-09958fba15cb\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf59846b-zc9sg" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.748938 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-84d4cc4998-msdhj" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.762862 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7fbf59846b-zc9sg" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.907500 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="358682a8-6b18-49ee-bd7d-d9eab2b22ee0" path="/var/lib/kubelet/pods/358682a8-6b18-49ee-bd7d-d9eab2b22ee0/volumes" Nov 24 07:07:28 crc kubenswrapper[4809]: I1124 07:07:28.908387 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3ac0673-5554-4ac6-a16e-b08631fdbf94" path="/var/lib/kubelet/pods/f3ac0673-5554-4ac6-a16e-b08631fdbf94/volumes" Nov 24 07:07:29 crc kubenswrapper[4809]: I1124 07:07:29.214957 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-84d4cc4998-msdhj"] Nov 24 07:07:29 crc kubenswrapper[4809]: I1124 07:07:29.252609 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fbf59846b-zc9sg"] Nov 24 07:07:29 crc kubenswrapper[4809]: I1124 07:07:29.546894 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-84d4cc4998-msdhj" event={"ID":"3b7a5653-eaf1-409a-a343-1e1ffb898b6b","Type":"ContainerStarted","Data":"39cc4174c422e593ec4ba666350e831e2c77c4aba1d281ab2d578f7d5baa0921"} Nov 24 07:07:29 crc kubenswrapper[4809]: I1124 07:07:29.546941 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-84d4cc4998-msdhj" event={"ID":"3b7a5653-eaf1-409a-a343-1e1ffb898b6b","Type":"ContainerStarted","Data":"9b8cbf166cc325e2d5c06d4a2ff75c4b06803d3e9e1d3e769c55c9d90ac6f0f7"} Nov 24 07:07:29 crc kubenswrapper[4809]: I1124 07:07:29.547274 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-84d4cc4998-msdhj" Nov 24 07:07:29 crc kubenswrapper[4809]: I1124 07:07:29.548114 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7fbf59846b-zc9sg" event={"ID":"4a48066b-1019-4c99-9c1e-09958fba15cb","Type":"ContainerStarted","Data":"958995f56bd9dd7cc728e6965300903d5f7476dd56bc71836de0493878ef44b9"} Nov 24 07:07:29 crc kubenswrapper[4809]: I1124 07:07:29.548144 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7fbf59846b-zc9sg" event={"ID":"4a48066b-1019-4c99-9c1e-09958fba15cb","Type":"ContainerStarted","Data":"5e5bed0ef7b9f94e512bcfad3019fc230241bf5fa652ae55d87f463071bc1e29"} Nov 24 07:07:29 crc kubenswrapper[4809]: I1124 07:07:29.548317 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7fbf59846b-zc9sg" Nov 24 07:07:29 crc kubenswrapper[4809]: I1124 07:07:29.558786 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-84d4cc4998-msdhj" Nov 24 07:07:29 crc kubenswrapper[4809]: I1124 07:07:29.580988 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-84d4cc4998-msdhj" podStartSLOduration=3.580946558 podStartE2EDuration="3.580946558s" podCreationTimestamp="2025-11-24 07:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:07:29.576261021 +0000 UTC m=+749.476852726" watchObservedRunningTime="2025-11-24 07:07:29.580946558 +0000 UTC m=+749.481538263" Nov 24 07:07:29 crc kubenswrapper[4809]: I1124 07:07:29.634044 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7fbf59846b-zc9sg" podStartSLOduration=3.634019038 podStartE2EDuration="3.634019038s" podCreationTimestamp="2025-11-24 07:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:07:29.630540743 +0000 UTC m=+749.531132458" watchObservedRunningTime="2025-11-24 07:07:29.634019038 +0000 UTC m=+749.534611043" Nov 24 07:07:30 crc kubenswrapper[4809]: I1124 07:07:30.011377 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7fbf59846b-zc9sg" Nov 24 07:07:33 crc kubenswrapper[4809]: I1124 07:07:33.485370 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-7d8556447d-spn4w" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.369052 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-b922n"] Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.370268 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-b922n" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.381953 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-z5gsd"] Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.384254 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.384653 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-zpdrx" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.392306 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-b922n"] Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.392483 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.399076 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.399752 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.448447 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-metrics-certs\") pod \"frr-k8s-z5gsd\" (UID: \"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b\") " pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.448522 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-frr-startup\") pod \"frr-k8s-z5gsd\" (UID: \"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b\") " pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.448554 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-frr-sockets\") pod \"frr-k8s-z5gsd\" (UID: \"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b\") " pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.448588 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-reloader\") pod \"frr-k8s-z5gsd\" (UID: \"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b\") " pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.448659 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bx8d\" (UniqueName: \"kubernetes.io/projected/ab5d28f0-1848-47f6-bdc8-444e258af818-kube-api-access-4bx8d\") pod \"frr-k8s-webhook-server-6998585d5-b922n\" (UID: \"ab5d28f0-1848-47f6-bdc8-444e258af818\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-b922n" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.448688 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-metrics\") pod \"frr-k8s-z5gsd\" (UID: \"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b\") " pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.448722 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-frr-conf\") pod \"frr-k8s-z5gsd\" (UID: \"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b\") " pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.448818 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ab5d28f0-1848-47f6-bdc8-444e258af818-cert\") pod \"frr-k8s-webhook-server-6998585d5-b922n\" (UID: \"ab5d28f0-1848-47f6-bdc8-444e258af818\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-b922n" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.448859 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4blqg\" (UniqueName: \"kubernetes.io/projected/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-kube-api-access-4blqg\") pod \"frr-k8s-z5gsd\" (UID: \"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b\") " pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.489557 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-fpksh"] Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.490663 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-fpksh" Nov 24 07:07:34 crc kubenswrapper[4809]: W1124 07:07:34.493005 4809 reflector.go:561] object-"metallb-system"/"metallb-memberlist": failed to list *v1.Secret: secrets "metallb-memberlist" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Nov 24 07:07:34 crc kubenswrapper[4809]: E1124 07:07:34.493046 4809 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"metallb-memberlist\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"metallb-memberlist\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 24 07:07:34 crc kubenswrapper[4809]: W1124 07:07:34.493232 4809 reflector.go:561] object-"metallb-system"/"speaker-certs-secret": failed to list *v1.Secret: secrets "speaker-certs-secret" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Nov 24 07:07:34 crc kubenswrapper[4809]: E1124 07:07:34.493252 4809 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"speaker-certs-secret\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"speaker-certs-secret\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 24 07:07:34 crc kubenswrapper[4809]: W1124 07:07:34.493573 4809 reflector.go:561] object-"metallb-system"/"speaker-dockercfg-cck9w": failed to list *v1.Secret: secrets "speaker-dockercfg-cck9w" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Nov 24 07:07:34 crc kubenswrapper[4809]: E1124 07:07:34.493594 4809 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"speaker-dockercfg-cck9w\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"speaker-dockercfg-cck9w\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 24 07:07:34 crc kubenswrapper[4809]: W1124 07:07:34.494577 4809 reflector.go:561] object-"metallb-system"/"metallb-excludel2": failed to list *v1.ConfigMap: configmaps "metallb-excludel2" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Nov 24 07:07:34 crc kubenswrapper[4809]: E1124 07:07:34.494601 4809 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"metallb-excludel2\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"metallb-excludel2\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.516860 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-8p4jp"] Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.517938 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-8p4jp" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.519978 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.548774 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-8p4jp"] Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.550010 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-frr-sockets\") pod \"frr-k8s-z5gsd\" (UID: \"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b\") " pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.550048 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-reloader\") pod \"frr-k8s-z5gsd\" (UID: \"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b\") " pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.550086 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bx8d\" (UniqueName: \"kubernetes.io/projected/ab5d28f0-1848-47f6-bdc8-444e258af818-kube-api-access-4bx8d\") pod \"frr-k8s-webhook-server-6998585d5-b922n\" (UID: \"ab5d28f0-1848-47f6-bdc8-444e258af818\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-b922n" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.550110 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-metrics\") pod \"frr-k8s-z5gsd\" (UID: \"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b\") " pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.550136 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/13297f30-8d99-42d0-9760-9f88705b691f-metrics-certs\") pod \"speaker-fpksh\" (UID: \"13297f30-8d99-42d0-9760-9f88705b691f\") " pod="metallb-system/speaker-fpksh" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.550159 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-frr-conf\") pod \"frr-k8s-z5gsd\" (UID: \"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b\") " pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.550184 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/13297f30-8d99-42d0-9760-9f88705b691f-memberlist\") pod \"speaker-fpksh\" (UID: \"13297f30-8d99-42d0-9760-9f88705b691f\") " pod="metallb-system/speaker-fpksh" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.550245 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ab5d28f0-1848-47f6-bdc8-444e258af818-cert\") pod \"frr-k8s-webhook-server-6998585d5-b922n\" (UID: \"ab5d28f0-1848-47f6-bdc8-444e258af818\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-b922n" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.550288 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/13297f30-8d99-42d0-9760-9f88705b691f-metallb-excludel2\") pod \"speaker-fpksh\" (UID: \"13297f30-8d99-42d0-9760-9f88705b691f\") " pod="metallb-system/speaker-fpksh" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.550316 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4blqg\" (UniqueName: \"kubernetes.io/projected/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-kube-api-access-4blqg\") pod \"frr-k8s-z5gsd\" (UID: \"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b\") " pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.550602 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-frr-sockets\") pod \"frr-k8s-z5gsd\" (UID: \"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b\") " pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.550625 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kj6z\" (UniqueName: \"kubernetes.io/projected/13297f30-8d99-42d0-9760-9f88705b691f-kube-api-access-5kj6z\") pod \"speaker-fpksh\" (UID: \"13297f30-8d99-42d0-9760-9f88705b691f\") " pod="metallb-system/speaker-fpksh" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.550658 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-frr-conf\") pod \"frr-k8s-z5gsd\" (UID: \"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b\") " pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.550705 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-reloader\") pod \"frr-k8s-z5gsd\" (UID: \"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b\") " pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.550731 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-metrics-certs\") pod \"frr-k8s-z5gsd\" (UID: \"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b\") " pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.550768 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-frr-startup\") pod \"frr-k8s-z5gsd\" (UID: \"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b\") " pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.550800 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-metrics\") pod \"frr-k8s-z5gsd\" (UID: \"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b\") " pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:34 crc kubenswrapper[4809]: E1124 07:07:34.550867 4809 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Nov 24 07:07:34 crc kubenswrapper[4809]: E1124 07:07:34.550923 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-metrics-certs podName:c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b nodeName:}" failed. No retries permitted until 2025-11-24 07:07:35.050897801 +0000 UTC m=+754.951489506 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-metrics-certs") pod "frr-k8s-z5gsd" (UID: "c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b") : secret "frr-k8s-certs-secret" not found Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.551716 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-frr-startup\") pod \"frr-k8s-z5gsd\" (UID: \"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b\") " pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.557148 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ab5d28f0-1848-47f6-bdc8-444e258af818-cert\") pod \"frr-k8s-webhook-server-6998585d5-b922n\" (UID: \"ab5d28f0-1848-47f6-bdc8-444e258af818\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-b922n" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.576798 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4blqg\" (UniqueName: \"kubernetes.io/projected/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-kube-api-access-4blqg\") pod \"frr-k8s-z5gsd\" (UID: \"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b\") " pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.602911 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bx8d\" (UniqueName: \"kubernetes.io/projected/ab5d28f0-1848-47f6-bdc8-444e258af818-kube-api-access-4bx8d\") pod \"frr-k8s-webhook-server-6998585d5-b922n\" (UID: \"ab5d28f0-1848-47f6-bdc8-444e258af818\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-b922n" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.651854 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/13297f30-8d99-42d0-9760-9f88705b691f-metallb-excludel2\") pod \"speaker-fpksh\" (UID: \"13297f30-8d99-42d0-9760-9f88705b691f\") " pod="metallb-system/speaker-fpksh" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.651897 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kj6z\" (UniqueName: \"kubernetes.io/projected/13297f30-8d99-42d0-9760-9f88705b691f-kube-api-access-5kj6z\") pod \"speaker-fpksh\" (UID: \"13297f30-8d99-42d0-9760-9f88705b691f\") " pod="metallb-system/speaker-fpksh" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.651917 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/00f3a099-57a2-4da1-bb22-12c48c359ab7-metrics-certs\") pod \"controller-6c7b4b5f48-8p4jp\" (UID: \"00f3a099-57a2-4da1-bb22-12c48c359ab7\") " pod="metallb-system/controller-6c7b4b5f48-8p4jp" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.652268 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws4m4\" (UniqueName: \"kubernetes.io/projected/00f3a099-57a2-4da1-bb22-12c48c359ab7-kube-api-access-ws4m4\") pod \"controller-6c7b4b5f48-8p4jp\" (UID: \"00f3a099-57a2-4da1-bb22-12c48c359ab7\") " pod="metallb-system/controller-6c7b4b5f48-8p4jp" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.652296 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/00f3a099-57a2-4da1-bb22-12c48c359ab7-cert\") pod \"controller-6c7b4b5f48-8p4jp\" (UID: \"00f3a099-57a2-4da1-bb22-12c48c359ab7\") " pod="metallb-system/controller-6c7b4b5f48-8p4jp" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.652349 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/13297f30-8d99-42d0-9760-9f88705b691f-metrics-certs\") pod \"speaker-fpksh\" (UID: \"13297f30-8d99-42d0-9760-9f88705b691f\") " pod="metallb-system/speaker-fpksh" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.652388 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/13297f30-8d99-42d0-9760-9f88705b691f-memberlist\") pod \"speaker-fpksh\" (UID: \"13297f30-8d99-42d0-9760-9f88705b691f\") " pod="metallb-system/speaker-fpksh" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.669363 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kj6z\" (UniqueName: \"kubernetes.io/projected/13297f30-8d99-42d0-9760-9f88705b691f-kube-api-access-5kj6z\") pod \"speaker-fpksh\" (UID: \"13297f30-8d99-42d0-9760-9f88705b691f\") " pod="metallb-system/speaker-fpksh" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.689066 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-b922n" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.754096 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/00f3a099-57a2-4da1-bb22-12c48c359ab7-metrics-certs\") pod \"controller-6c7b4b5f48-8p4jp\" (UID: \"00f3a099-57a2-4da1-bb22-12c48c359ab7\") " pod="metallb-system/controller-6c7b4b5f48-8p4jp" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.754227 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws4m4\" (UniqueName: \"kubernetes.io/projected/00f3a099-57a2-4da1-bb22-12c48c359ab7-kube-api-access-ws4m4\") pod \"controller-6c7b4b5f48-8p4jp\" (UID: \"00f3a099-57a2-4da1-bb22-12c48c359ab7\") " pod="metallb-system/controller-6c7b4b5f48-8p4jp" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.754255 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/00f3a099-57a2-4da1-bb22-12c48c359ab7-cert\") pod \"controller-6c7b4b5f48-8p4jp\" (UID: \"00f3a099-57a2-4da1-bb22-12c48c359ab7\") " pod="metallb-system/controller-6c7b4b5f48-8p4jp" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.757380 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.757900 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/00f3a099-57a2-4da1-bb22-12c48c359ab7-metrics-certs\") pod \"controller-6c7b4b5f48-8p4jp\" (UID: \"00f3a099-57a2-4da1-bb22-12c48c359ab7\") " pod="metallb-system/controller-6c7b4b5f48-8p4jp" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.776198 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/00f3a099-57a2-4da1-bb22-12c48c359ab7-cert\") pod \"controller-6c7b4b5f48-8p4jp\" (UID: \"00f3a099-57a2-4da1-bb22-12c48c359ab7\") " pod="metallb-system/controller-6c7b4b5f48-8p4jp" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.786829 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws4m4\" (UniqueName: \"kubernetes.io/projected/00f3a099-57a2-4da1-bb22-12c48c359ab7-kube-api-access-ws4m4\") pod \"controller-6c7b4b5f48-8p4jp\" (UID: \"00f3a099-57a2-4da1-bb22-12c48c359ab7\") " pod="metallb-system/controller-6c7b4b5f48-8p4jp" Nov 24 07:07:34 crc kubenswrapper[4809]: I1124 07:07:34.831858 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-8p4jp" Nov 24 07:07:35 crc kubenswrapper[4809]: I1124 07:07:35.060581 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-metrics-certs\") pod \"frr-k8s-z5gsd\" (UID: \"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b\") " pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:35 crc kubenswrapper[4809]: I1124 07:07:35.064336 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b-metrics-certs\") pod \"frr-k8s-z5gsd\" (UID: \"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b\") " pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:35 crc kubenswrapper[4809]: I1124 07:07:35.091935 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-b922n"] Nov 24 07:07:35 crc kubenswrapper[4809]: I1124 07:07:35.256730 4809 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 07:07:35 crc kubenswrapper[4809]: I1124 07:07:35.259856 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-8p4jp"] Nov 24 07:07:35 crc kubenswrapper[4809]: W1124 07:07:35.266763 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00f3a099_57a2_4da1_bb22_12c48c359ab7.slice/crio-067aab9c808183c0c4f191b9b39ad2b07defcb8cc29fd707179c87292648fbd0 WatchSource:0}: Error finding container 067aab9c808183c0c4f191b9b39ad2b07defcb8cc29fd707179c87292648fbd0: Status 404 returned error can't find the container with id 067aab9c808183c0c4f191b9b39ad2b07defcb8cc29fd707179c87292648fbd0 Nov 24 07:07:35 crc kubenswrapper[4809]: I1124 07:07:35.307145 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:35 crc kubenswrapper[4809]: I1124 07:07:35.369032 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 24 07:07:35 crc kubenswrapper[4809]: I1124 07:07:35.377519 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/13297f30-8d99-42d0-9760-9f88705b691f-metrics-certs\") pod \"speaker-fpksh\" (UID: \"13297f30-8d99-42d0-9760-9f88705b691f\") " pod="metallb-system/speaker-fpksh" Nov 24 07:07:35 crc kubenswrapper[4809]: I1124 07:07:35.589656 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-z5gsd" event={"ID":"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b","Type":"ContainerStarted","Data":"d83ead0db59b8ba0e2ec1d9b816010227b03c55f59f56f73d55165ba485ba2b6"} Nov 24 07:07:35 crc kubenswrapper[4809]: I1124 07:07:35.591481 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-8p4jp" event={"ID":"00f3a099-57a2-4da1-bb22-12c48c359ab7","Type":"ContainerStarted","Data":"948a71b298b4eb7c87bdcab45d7240f93b6e0cb16abee6a8fecb5ace152dc900"} Nov 24 07:07:35 crc kubenswrapper[4809]: I1124 07:07:35.591535 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-8p4jp" event={"ID":"00f3a099-57a2-4da1-bb22-12c48c359ab7","Type":"ContainerStarted","Data":"821bfcc3bce5c722be08569b6def6e3f9ea8b2225a98b227e8b54e12ef42b0e0"} Nov 24 07:07:35 crc kubenswrapper[4809]: I1124 07:07:35.591560 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-8p4jp" event={"ID":"00f3a099-57a2-4da1-bb22-12c48c359ab7","Type":"ContainerStarted","Data":"067aab9c808183c0c4f191b9b39ad2b07defcb8cc29fd707179c87292648fbd0"} Nov 24 07:07:35 crc kubenswrapper[4809]: I1124 07:07:35.591576 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-8p4jp" Nov 24 07:07:35 crc kubenswrapper[4809]: I1124 07:07:35.592493 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-b922n" event={"ID":"ab5d28f0-1848-47f6-bdc8-444e258af818","Type":"ContainerStarted","Data":"defa7c7e0a1c780fc0715509ecbd2f83d25aec4e2ae1a8d54003b06a64f7c814"} Nov 24 07:07:35 crc kubenswrapper[4809]: I1124 07:07:35.608087 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-8p4jp" podStartSLOduration=1.6080683919999998 podStartE2EDuration="1.608068392s" podCreationTimestamp="2025-11-24 07:07:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:07:35.604237368 +0000 UTC m=+755.504829083" watchObservedRunningTime="2025-11-24 07:07:35.608068392 +0000 UTC m=+755.508660097" Nov 24 07:07:35 crc kubenswrapper[4809]: E1124 07:07:35.652924 4809 configmap.go:193] Couldn't get configMap metallb-system/metallb-excludel2: failed to sync configmap cache: timed out waiting for the condition Nov 24 07:07:35 crc kubenswrapper[4809]: E1124 07:07:35.652956 4809 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: failed to sync secret cache: timed out waiting for the condition Nov 24 07:07:35 crc kubenswrapper[4809]: E1124 07:07:35.653034 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/13297f30-8d99-42d0-9760-9f88705b691f-metallb-excludel2 podName:13297f30-8d99-42d0-9760-9f88705b691f nodeName:}" failed. No retries permitted until 2025-11-24 07:07:36.153015432 +0000 UTC m=+756.053607137 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metallb-excludel2" (UniqueName: "kubernetes.io/configmap/13297f30-8d99-42d0-9760-9f88705b691f-metallb-excludel2") pod "speaker-fpksh" (UID: "13297f30-8d99-42d0-9760-9f88705b691f") : failed to sync configmap cache: timed out waiting for the condition Nov 24 07:07:35 crc kubenswrapper[4809]: E1124 07:07:35.653082 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/13297f30-8d99-42d0-9760-9f88705b691f-memberlist podName:13297f30-8d99-42d0-9760-9f88705b691f nodeName:}" failed. No retries permitted until 2025-11-24 07:07:36.153054223 +0000 UTC m=+756.053645968 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/13297f30-8d99-42d0-9760-9f88705b691f-memberlist") pod "speaker-fpksh" (UID: "13297f30-8d99-42d0-9760-9f88705b691f") : failed to sync secret cache: timed out waiting for the condition Nov 24 07:07:35 crc kubenswrapper[4809]: I1124 07:07:35.807785 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-cck9w" Nov 24 07:07:35 crc kubenswrapper[4809]: I1124 07:07:35.849899 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 24 07:07:36 crc kubenswrapper[4809]: I1124 07:07:36.069072 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 24 07:07:36 crc kubenswrapper[4809]: I1124 07:07:36.174784 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/13297f30-8d99-42d0-9760-9f88705b691f-metallb-excludel2\") pod \"speaker-fpksh\" (UID: \"13297f30-8d99-42d0-9760-9f88705b691f\") " pod="metallb-system/speaker-fpksh" Nov 24 07:07:36 crc kubenswrapper[4809]: I1124 07:07:36.174941 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/13297f30-8d99-42d0-9760-9f88705b691f-memberlist\") pod \"speaker-fpksh\" (UID: \"13297f30-8d99-42d0-9760-9f88705b691f\") " pod="metallb-system/speaker-fpksh" Nov 24 07:07:36 crc kubenswrapper[4809]: I1124 07:07:36.175536 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/13297f30-8d99-42d0-9760-9f88705b691f-metallb-excludel2\") pod \"speaker-fpksh\" (UID: \"13297f30-8d99-42d0-9760-9f88705b691f\") " pod="metallb-system/speaker-fpksh" Nov 24 07:07:36 crc kubenswrapper[4809]: I1124 07:07:36.182490 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/13297f30-8d99-42d0-9760-9f88705b691f-memberlist\") pod \"speaker-fpksh\" (UID: \"13297f30-8d99-42d0-9760-9f88705b691f\") " pod="metallb-system/speaker-fpksh" Nov 24 07:07:36 crc kubenswrapper[4809]: I1124 07:07:36.303077 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-fpksh" Nov 24 07:07:36 crc kubenswrapper[4809]: W1124 07:07:36.319042 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13297f30_8d99_42d0_9760_9f88705b691f.slice/crio-05f8a5c5c303702357756960801e0775903b86d058b6df7f05e0b4f2aa4aa401 WatchSource:0}: Error finding container 05f8a5c5c303702357756960801e0775903b86d058b6df7f05e0b4f2aa4aa401: Status 404 returned error can't find the container with id 05f8a5c5c303702357756960801e0775903b86d058b6df7f05e0b4f2aa4aa401 Nov 24 07:07:36 crc kubenswrapper[4809]: I1124 07:07:36.603537 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-fpksh" event={"ID":"13297f30-8d99-42d0-9760-9f88705b691f","Type":"ContainerStarted","Data":"eecbdd5918dac0d8693d1f1e28ac9e19cfe9f0b89329e4ea61ad1470a407ef8f"} Nov 24 07:07:36 crc kubenswrapper[4809]: I1124 07:07:36.603772 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-fpksh" event={"ID":"13297f30-8d99-42d0-9760-9f88705b691f","Type":"ContainerStarted","Data":"05f8a5c5c303702357756960801e0775903b86d058b6df7f05e0b4f2aa4aa401"} Nov 24 07:07:37 crc kubenswrapper[4809]: I1124 07:07:37.612030 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-fpksh" event={"ID":"13297f30-8d99-42d0-9760-9f88705b691f","Type":"ContainerStarted","Data":"366c0d463132e75f612581257a1bdf2c7d9e1e71be1492a88016f0660608c31c"} Nov 24 07:07:37 crc kubenswrapper[4809]: I1124 07:07:37.612818 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-fpksh" Nov 24 07:07:40 crc kubenswrapper[4809]: I1124 07:07:40.915215 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-fpksh" podStartSLOduration=6.915197383 podStartE2EDuration="6.915197383s" podCreationTimestamp="2025-11-24 07:07:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:07:37.62941767 +0000 UTC m=+757.530009375" watchObservedRunningTime="2025-11-24 07:07:40.915197383 +0000 UTC m=+760.815789078" Nov 24 07:07:41 crc kubenswrapper[4809]: I1124 07:07:41.643385 4809 generic.go:334] "Generic (PLEG): container finished" podID="c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b" containerID="f66fb936b10c89628f367371dc4c649ef0643a08ab9fdef8b034cf4f371a0054" exitCode=0 Nov 24 07:07:41 crc kubenswrapper[4809]: I1124 07:07:41.643604 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-z5gsd" event={"ID":"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b","Type":"ContainerDied","Data":"f66fb936b10c89628f367371dc4c649ef0643a08ab9fdef8b034cf4f371a0054"} Nov 24 07:07:41 crc kubenswrapper[4809]: I1124 07:07:41.645715 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-b922n" event={"ID":"ab5d28f0-1848-47f6-bdc8-444e258af818","Type":"ContainerStarted","Data":"b8ab2d343eb175ff47f64eda31671a48a0ded1af5af9a81d6339781d9d9a0083"} Nov 24 07:07:41 crc kubenswrapper[4809]: I1124 07:07:41.646263 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-b922n" Nov 24 07:07:41 crc kubenswrapper[4809]: I1124 07:07:41.691531 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-b922n" podStartSLOduration=1.338726703 podStartE2EDuration="7.691514044s" podCreationTimestamp="2025-11-24 07:07:34 +0000 UTC" firstStartedPulling="2025-11-24 07:07:35.098641161 +0000 UTC m=+754.999232856" lastFinishedPulling="2025-11-24 07:07:41.451428492 +0000 UTC m=+761.352020197" observedRunningTime="2025-11-24 07:07:41.688922584 +0000 UTC m=+761.589514289" watchObservedRunningTime="2025-11-24 07:07:41.691514044 +0000 UTC m=+761.592105749" Nov 24 07:07:42 crc kubenswrapper[4809]: I1124 07:07:42.654382 4809 generic.go:334] "Generic (PLEG): container finished" podID="c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b" containerID="8f54c63ddf335d204d981e81af843d8a0eccd6f25d0aaad4d50ee40ef2c7051b" exitCode=0 Nov 24 07:07:42 crc kubenswrapper[4809]: I1124 07:07:42.654436 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-z5gsd" event={"ID":"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b","Type":"ContainerDied","Data":"8f54c63ddf335d204d981e81af843d8a0eccd6f25d0aaad4d50ee40ef2c7051b"} Nov 24 07:07:43 crc kubenswrapper[4809]: I1124 07:07:43.664276 4809 generic.go:334] "Generic (PLEG): container finished" podID="c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b" containerID="835fe2a4ece7c9fc00e94e4e9c9f08403d81b2f8b8f6cd8e2897bc6ad59d8cdb" exitCode=0 Nov 24 07:07:43 crc kubenswrapper[4809]: I1124 07:07:43.664335 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-z5gsd" event={"ID":"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b","Type":"ContainerDied","Data":"835fe2a4ece7c9fc00e94e4e9c9f08403d81b2f8b8f6cd8e2897bc6ad59d8cdb"} Nov 24 07:07:44 crc kubenswrapper[4809]: I1124 07:07:44.673307 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-z5gsd" event={"ID":"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b","Type":"ContainerStarted","Data":"c1fe2e123a3a9895fee28fd6dd4b7ef5302625a01e168a5abdea6fc8f55162be"} Nov 24 07:07:44 crc kubenswrapper[4809]: I1124 07:07:44.673591 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-z5gsd" event={"ID":"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b","Type":"ContainerStarted","Data":"9d6f755948275ae282b67c0b10ee4e4041504a306715abf74c19fcfab3089bdd"} Nov 24 07:07:44 crc kubenswrapper[4809]: I1124 07:07:44.673602 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-z5gsd" event={"ID":"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b","Type":"ContainerStarted","Data":"bd6b6ba6a2b1c11004af6faf2e52706132a8623a7ffc1b8bef8249b28f9fdc95"} Nov 24 07:07:44 crc kubenswrapper[4809]: I1124 07:07:44.673614 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-z5gsd" event={"ID":"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b","Type":"ContainerStarted","Data":"fc90107584be65dae8d5cb3cf295ed01bf1895eb1cc3c1406859e9888f1aee1c"} Nov 24 07:07:45 crc kubenswrapper[4809]: I1124 07:07:45.687124 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-z5gsd" event={"ID":"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b","Type":"ContainerStarted","Data":"4ae3115cc39a82a23d283fbab2120056f6e2e6688e8d8396b1fd60fbf42d1014"} Nov 24 07:07:45 crc kubenswrapper[4809]: I1124 07:07:45.687793 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-z5gsd" event={"ID":"c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b","Type":"ContainerStarted","Data":"7630d3ee4252a59e690a666773e50b815eff0d487da8ec86e590d6101a128310"} Nov 24 07:07:45 crc kubenswrapper[4809]: I1124 07:07:45.687894 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:45 crc kubenswrapper[4809]: I1124 07:07:45.709278 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-z5gsd" podStartSLOduration=5.677481724 podStartE2EDuration="11.709240074s" podCreationTimestamp="2025-11-24 07:07:34 +0000 UTC" firstStartedPulling="2025-11-24 07:07:35.43437825 +0000 UTC m=+755.334969955" lastFinishedPulling="2025-11-24 07:07:41.4661366 +0000 UTC m=+761.366728305" observedRunningTime="2025-11-24 07:07:45.706871931 +0000 UTC m=+765.607463676" watchObservedRunningTime="2025-11-24 07:07:45.709240074 +0000 UTC m=+765.609831799" Nov 24 07:07:46 crc kubenswrapper[4809]: I1124 07:07:46.332643 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-fpksh" Nov 24 07:07:48 crc kubenswrapper[4809]: I1124 07:07:48.043588 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:07:48 crc kubenswrapper[4809]: I1124 07:07:48.044125 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:07:48 crc kubenswrapper[4809]: I1124 07:07:48.044200 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 07:07:48 crc kubenswrapper[4809]: I1124 07:07:48.045153 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"672cd2d0a97571eba82aabde6a902d139a7242488c7021eea3d01afc3b311c16"} pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:07:48 crc kubenswrapper[4809]: I1124 07:07:48.045238 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" containerID="cri-o://672cd2d0a97571eba82aabde6a902d139a7242488c7021eea3d01afc3b311c16" gracePeriod=600 Nov 24 07:07:48 crc kubenswrapper[4809]: I1124 07:07:48.708506 4809 generic.go:334] "Generic (PLEG): container finished" podID="be8050d5-1aea-453e-a06e-87702b0e856b" containerID="672cd2d0a97571eba82aabde6a902d139a7242488c7021eea3d01afc3b311c16" exitCode=0 Nov 24 07:07:48 crc kubenswrapper[4809]: I1124 07:07:48.708600 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerDied","Data":"672cd2d0a97571eba82aabde6a902d139a7242488c7021eea3d01afc3b311c16"} Nov 24 07:07:48 crc kubenswrapper[4809]: I1124 07:07:48.708924 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerStarted","Data":"5bbd5cb6a3a9e9de8c1c6f89b55f7b7f733fb4a270390a494ec13ae9d9f8fbca"} Nov 24 07:07:48 crc kubenswrapper[4809]: I1124 07:07:48.708993 4809 scope.go:117] "RemoveContainer" containerID="24354d43179411329fa3f10b3bcaad9ac5fbb3505999caa72b922a82781a0f78" Nov 24 07:07:49 crc kubenswrapper[4809]: I1124 07:07:49.547951 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-tjw95"] Nov 24 07:07:49 crc kubenswrapper[4809]: I1124 07:07:49.549214 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tjw95" Nov 24 07:07:49 crc kubenswrapper[4809]: I1124 07:07:49.552446 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 24 07:07:49 crc kubenswrapper[4809]: I1124 07:07:49.566374 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 24 07:07:49 crc kubenswrapper[4809]: I1124 07:07:49.567024 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-tjw95"] Nov 24 07:07:49 crc kubenswrapper[4809]: I1124 07:07:49.670815 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smmhm\" (UniqueName: \"kubernetes.io/projected/aa235167-881e-448a-aac4-126b45a73e79-kube-api-access-smmhm\") pod \"openstack-operator-index-tjw95\" (UID: \"aa235167-881e-448a-aac4-126b45a73e79\") " pod="openstack-operators/openstack-operator-index-tjw95" Nov 24 07:07:49 crc kubenswrapper[4809]: I1124 07:07:49.772270 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smmhm\" (UniqueName: \"kubernetes.io/projected/aa235167-881e-448a-aac4-126b45a73e79-kube-api-access-smmhm\") pod \"openstack-operator-index-tjw95\" (UID: \"aa235167-881e-448a-aac4-126b45a73e79\") " pod="openstack-operators/openstack-operator-index-tjw95" Nov 24 07:07:49 crc kubenswrapper[4809]: I1124 07:07:49.788747 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smmhm\" (UniqueName: \"kubernetes.io/projected/aa235167-881e-448a-aac4-126b45a73e79-kube-api-access-smmhm\") pod \"openstack-operator-index-tjw95\" (UID: \"aa235167-881e-448a-aac4-126b45a73e79\") " pod="openstack-operators/openstack-operator-index-tjw95" Nov 24 07:07:49 crc kubenswrapper[4809]: I1124 07:07:49.870198 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tjw95" Nov 24 07:07:50 crc kubenswrapper[4809]: I1124 07:07:50.308819 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:50 crc kubenswrapper[4809]: I1124 07:07:50.348706 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:50 crc kubenswrapper[4809]: I1124 07:07:50.357226 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-tjw95"] Nov 24 07:07:50 crc kubenswrapper[4809]: I1124 07:07:50.723756 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tjw95" event={"ID":"aa235167-881e-448a-aac4-126b45a73e79","Type":"ContainerStarted","Data":"d4b8a8d38bbb4e9b3f5ec4700b46f2714ff654e85a6ea92acc2f75526c2df977"} Nov 24 07:07:52 crc kubenswrapper[4809]: I1124 07:07:52.738826 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tjw95" event={"ID":"aa235167-881e-448a-aac4-126b45a73e79","Type":"ContainerStarted","Data":"057f921b7c3d40133301e9222c17f4bc63fe6284d8324b8705386d8a0cf3127f"} Nov 24 07:07:52 crc kubenswrapper[4809]: I1124 07:07:52.762830 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-tjw95" podStartSLOduration=1.7576821759999999 podStartE2EDuration="3.762807665s" podCreationTimestamp="2025-11-24 07:07:49 +0000 UTC" firstStartedPulling="2025-11-24 07:07:50.360932533 +0000 UTC m=+770.261524238" lastFinishedPulling="2025-11-24 07:07:52.366057992 +0000 UTC m=+772.266649727" observedRunningTime="2025-11-24 07:07:52.753700958 +0000 UTC m=+772.654292693" watchObservedRunningTime="2025-11-24 07:07:52.762807665 +0000 UTC m=+772.663399380" Nov 24 07:07:52 crc kubenswrapper[4809]: I1124 07:07:52.915231 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-tjw95"] Nov 24 07:07:53 crc kubenswrapper[4809]: I1124 07:07:53.528222 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-mhj4l"] Nov 24 07:07:53 crc kubenswrapper[4809]: I1124 07:07:53.530215 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mhj4l" Nov 24 07:07:53 crc kubenswrapper[4809]: I1124 07:07:53.533815 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-cwqb6" Nov 24 07:07:53 crc kubenswrapper[4809]: I1124 07:07:53.570202 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-mhj4l"] Nov 24 07:07:53 crc kubenswrapper[4809]: I1124 07:07:53.624056 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmf24\" (UniqueName: \"kubernetes.io/projected/bb31150b-b8eb-4340-ba08-384de196289a-kube-api-access-rmf24\") pod \"openstack-operator-index-mhj4l\" (UID: \"bb31150b-b8eb-4340-ba08-384de196289a\") " pod="openstack-operators/openstack-operator-index-mhj4l" Nov 24 07:07:53 crc kubenswrapper[4809]: I1124 07:07:53.725576 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmf24\" (UniqueName: \"kubernetes.io/projected/bb31150b-b8eb-4340-ba08-384de196289a-kube-api-access-rmf24\") pod \"openstack-operator-index-mhj4l\" (UID: \"bb31150b-b8eb-4340-ba08-384de196289a\") " pod="openstack-operators/openstack-operator-index-mhj4l" Nov 24 07:07:53 crc kubenswrapper[4809]: I1124 07:07:53.752357 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmf24\" (UniqueName: \"kubernetes.io/projected/bb31150b-b8eb-4340-ba08-384de196289a-kube-api-access-rmf24\") pod \"openstack-operator-index-mhj4l\" (UID: \"bb31150b-b8eb-4340-ba08-384de196289a\") " pod="openstack-operators/openstack-operator-index-mhj4l" Nov 24 07:07:53 crc kubenswrapper[4809]: I1124 07:07:53.886617 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mhj4l" Nov 24 07:07:54 crc kubenswrapper[4809]: I1124 07:07:54.393901 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-mhj4l"] Nov 24 07:07:54 crc kubenswrapper[4809]: W1124 07:07:54.405937 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb31150b_b8eb_4340_ba08_384de196289a.slice/crio-0e9e09c5a0b89538da0190de72201b0c5d24701916fa51ef3a336f4eb9ddca5e WatchSource:0}: Error finding container 0e9e09c5a0b89538da0190de72201b0c5d24701916fa51ef3a336f4eb9ddca5e: Status 404 returned error can't find the container with id 0e9e09c5a0b89538da0190de72201b0c5d24701916fa51ef3a336f4eb9ddca5e Nov 24 07:07:54 crc kubenswrapper[4809]: I1124 07:07:54.694910 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-b922n" Nov 24 07:07:54 crc kubenswrapper[4809]: I1124 07:07:54.753758 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mhj4l" event={"ID":"bb31150b-b8eb-4340-ba08-384de196289a","Type":"ContainerStarted","Data":"1e539677eed9b79bd5a66a6eac88335df5027f65b2f0de4939d4227176fca4bc"} Nov 24 07:07:54 crc kubenswrapper[4809]: I1124 07:07:54.753823 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mhj4l" event={"ID":"bb31150b-b8eb-4340-ba08-384de196289a","Type":"ContainerStarted","Data":"0e9e09c5a0b89538da0190de72201b0c5d24701916fa51ef3a336f4eb9ddca5e"} Nov 24 07:07:54 crc kubenswrapper[4809]: I1124 07:07:54.753821 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-tjw95" podUID="aa235167-881e-448a-aac4-126b45a73e79" containerName="registry-server" containerID="cri-o://057f921b7c3d40133301e9222c17f4bc63fe6284d8324b8705386d8a0cf3127f" gracePeriod=2 Nov 24 07:07:54 crc kubenswrapper[4809]: I1124 07:07:54.776507 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-mhj4l" podStartSLOduration=1.720096806 podStartE2EDuration="1.776490926s" podCreationTimestamp="2025-11-24 07:07:53 +0000 UTC" firstStartedPulling="2025-11-24 07:07:54.411877815 +0000 UTC m=+774.312469570" lastFinishedPulling="2025-11-24 07:07:54.468271945 +0000 UTC m=+774.368863690" observedRunningTime="2025-11-24 07:07:54.773589038 +0000 UTC m=+774.674180793" watchObservedRunningTime="2025-11-24 07:07:54.776490926 +0000 UTC m=+774.677082631" Nov 24 07:07:54 crc kubenswrapper[4809]: I1124 07:07:54.836495 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-8p4jp" Nov 24 07:07:55 crc kubenswrapper[4809]: I1124 07:07:55.200525 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tjw95" Nov 24 07:07:55 crc kubenswrapper[4809]: I1124 07:07:55.312566 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-z5gsd" Nov 24 07:07:55 crc kubenswrapper[4809]: I1124 07:07:55.351047 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smmhm\" (UniqueName: \"kubernetes.io/projected/aa235167-881e-448a-aac4-126b45a73e79-kube-api-access-smmhm\") pod \"aa235167-881e-448a-aac4-126b45a73e79\" (UID: \"aa235167-881e-448a-aac4-126b45a73e79\") " Nov 24 07:07:55 crc kubenswrapper[4809]: I1124 07:07:55.356882 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa235167-881e-448a-aac4-126b45a73e79-kube-api-access-smmhm" (OuterVolumeSpecName: "kube-api-access-smmhm") pod "aa235167-881e-448a-aac4-126b45a73e79" (UID: "aa235167-881e-448a-aac4-126b45a73e79"). InnerVolumeSpecName "kube-api-access-smmhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:07:55 crc kubenswrapper[4809]: I1124 07:07:55.453190 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smmhm\" (UniqueName: \"kubernetes.io/projected/aa235167-881e-448a-aac4-126b45a73e79-kube-api-access-smmhm\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:55 crc kubenswrapper[4809]: I1124 07:07:55.761811 4809 generic.go:334] "Generic (PLEG): container finished" podID="aa235167-881e-448a-aac4-126b45a73e79" containerID="057f921b7c3d40133301e9222c17f4bc63fe6284d8324b8705386d8a0cf3127f" exitCode=0 Nov 24 07:07:55 crc kubenswrapper[4809]: I1124 07:07:55.761879 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tjw95" Nov 24 07:07:55 crc kubenswrapper[4809]: I1124 07:07:55.761868 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tjw95" event={"ID":"aa235167-881e-448a-aac4-126b45a73e79","Type":"ContainerDied","Data":"057f921b7c3d40133301e9222c17f4bc63fe6284d8324b8705386d8a0cf3127f"} Nov 24 07:07:55 crc kubenswrapper[4809]: I1124 07:07:55.762341 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tjw95" event={"ID":"aa235167-881e-448a-aac4-126b45a73e79","Type":"ContainerDied","Data":"d4b8a8d38bbb4e9b3f5ec4700b46f2714ff654e85a6ea92acc2f75526c2df977"} Nov 24 07:07:55 crc kubenswrapper[4809]: I1124 07:07:55.762396 4809 scope.go:117] "RemoveContainer" containerID="057f921b7c3d40133301e9222c17f4bc63fe6284d8324b8705386d8a0cf3127f" Nov 24 07:07:55 crc kubenswrapper[4809]: I1124 07:07:55.782018 4809 scope.go:117] "RemoveContainer" containerID="057f921b7c3d40133301e9222c17f4bc63fe6284d8324b8705386d8a0cf3127f" Nov 24 07:07:55 crc kubenswrapper[4809]: E1124 07:07:55.782673 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"057f921b7c3d40133301e9222c17f4bc63fe6284d8324b8705386d8a0cf3127f\": container with ID starting with 057f921b7c3d40133301e9222c17f4bc63fe6284d8324b8705386d8a0cf3127f not found: ID does not exist" containerID="057f921b7c3d40133301e9222c17f4bc63fe6284d8324b8705386d8a0cf3127f" Nov 24 07:07:55 crc kubenswrapper[4809]: I1124 07:07:55.782720 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"057f921b7c3d40133301e9222c17f4bc63fe6284d8324b8705386d8a0cf3127f"} err="failed to get container status \"057f921b7c3d40133301e9222c17f4bc63fe6284d8324b8705386d8a0cf3127f\": rpc error: code = NotFound desc = could not find container \"057f921b7c3d40133301e9222c17f4bc63fe6284d8324b8705386d8a0cf3127f\": container with ID starting with 057f921b7c3d40133301e9222c17f4bc63fe6284d8324b8705386d8a0cf3127f not found: ID does not exist" Nov 24 07:07:55 crc kubenswrapper[4809]: I1124 07:07:55.793727 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-tjw95"] Nov 24 07:07:55 crc kubenswrapper[4809]: I1124 07:07:55.803066 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-tjw95"] Nov 24 07:07:56 crc kubenswrapper[4809]: I1124 07:07:56.900433 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa235167-881e-448a-aac4-126b45a73e79" path="/var/lib/kubelet/pods/aa235167-881e-448a-aac4-126b45a73e79/volumes" Nov 24 07:07:57 crc kubenswrapper[4809]: I1124 07:07:57.728124 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dp25g"] Nov 24 07:07:57 crc kubenswrapper[4809]: E1124 07:07:57.728402 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa235167-881e-448a-aac4-126b45a73e79" containerName="registry-server" Nov 24 07:07:57 crc kubenswrapper[4809]: I1124 07:07:57.728418 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa235167-881e-448a-aac4-126b45a73e79" containerName="registry-server" Nov 24 07:07:57 crc kubenswrapper[4809]: I1124 07:07:57.728602 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa235167-881e-448a-aac4-126b45a73e79" containerName="registry-server" Nov 24 07:07:57 crc kubenswrapper[4809]: I1124 07:07:57.729678 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dp25g" Nov 24 07:07:57 crc kubenswrapper[4809]: I1124 07:07:57.747906 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dp25g"] Nov 24 07:07:57 crc kubenswrapper[4809]: I1124 07:07:57.892474 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80dfda6c-ae33-4dbd-9750-67079eeacffc-catalog-content\") pod \"community-operators-dp25g\" (UID: \"80dfda6c-ae33-4dbd-9750-67079eeacffc\") " pod="openshift-marketplace/community-operators-dp25g" Nov 24 07:07:57 crc kubenswrapper[4809]: I1124 07:07:57.892692 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57vlq\" (UniqueName: \"kubernetes.io/projected/80dfda6c-ae33-4dbd-9750-67079eeacffc-kube-api-access-57vlq\") pod \"community-operators-dp25g\" (UID: \"80dfda6c-ae33-4dbd-9750-67079eeacffc\") " pod="openshift-marketplace/community-operators-dp25g" Nov 24 07:07:57 crc kubenswrapper[4809]: I1124 07:07:57.892830 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80dfda6c-ae33-4dbd-9750-67079eeacffc-utilities\") pod \"community-operators-dp25g\" (UID: \"80dfda6c-ae33-4dbd-9750-67079eeacffc\") " pod="openshift-marketplace/community-operators-dp25g" Nov 24 07:07:57 crc kubenswrapper[4809]: I1124 07:07:57.994616 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57vlq\" (UniqueName: \"kubernetes.io/projected/80dfda6c-ae33-4dbd-9750-67079eeacffc-kube-api-access-57vlq\") pod \"community-operators-dp25g\" (UID: \"80dfda6c-ae33-4dbd-9750-67079eeacffc\") " pod="openshift-marketplace/community-operators-dp25g" Nov 24 07:07:57 crc kubenswrapper[4809]: I1124 07:07:57.994711 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80dfda6c-ae33-4dbd-9750-67079eeacffc-utilities\") pod \"community-operators-dp25g\" (UID: \"80dfda6c-ae33-4dbd-9750-67079eeacffc\") " pod="openshift-marketplace/community-operators-dp25g" Nov 24 07:07:57 crc kubenswrapper[4809]: I1124 07:07:57.994780 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80dfda6c-ae33-4dbd-9750-67079eeacffc-catalog-content\") pod \"community-operators-dp25g\" (UID: \"80dfda6c-ae33-4dbd-9750-67079eeacffc\") " pod="openshift-marketplace/community-operators-dp25g" Nov 24 07:07:57 crc kubenswrapper[4809]: I1124 07:07:57.995257 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80dfda6c-ae33-4dbd-9750-67079eeacffc-utilities\") pod \"community-operators-dp25g\" (UID: \"80dfda6c-ae33-4dbd-9750-67079eeacffc\") " pod="openshift-marketplace/community-operators-dp25g" Nov 24 07:07:57 crc kubenswrapper[4809]: I1124 07:07:57.995282 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80dfda6c-ae33-4dbd-9750-67079eeacffc-catalog-content\") pod \"community-operators-dp25g\" (UID: \"80dfda6c-ae33-4dbd-9750-67079eeacffc\") " pod="openshift-marketplace/community-operators-dp25g" Nov 24 07:07:58 crc kubenswrapper[4809]: I1124 07:07:58.013455 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57vlq\" (UniqueName: \"kubernetes.io/projected/80dfda6c-ae33-4dbd-9750-67079eeacffc-kube-api-access-57vlq\") pod \"community-operators-dp25g\" (UID: \"80dfda6c-ae33-4dbd-9750-67079eeacffc\") " pod="openshift-marketplace/community-operators-dp25g" Nov 24 07:07:58 crc kubenswrapper[4809]: I1124 07:07:58.050246 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dp25g" Nov 24 07:07:58 crc kubenswrapper[4809]: I1124 07:07:58.513194 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8nfm5"] Nov 24 07:07:58 crc kubenswrapper[4809]: I1124 07:07:58.514987 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8nfm5" Nov 24 07:07:58 crc kubenswrapper[4809]: I1124 07:07:58.526186 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dp25g"] Nov 24 07:07:58 crc kubenswrapper[4809]: I1124 07:07:58.531094 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8nfm5"] Nov 24 07:07:58 crc kubenswrapper[4809]: I1124 07:07:58.603290 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e191709-50db-49f9-b360-4d0e6021caf7-catalog-content\") pod \"redhat-marketplace-8nfm5\" (UID: \"4e191709-50db-49f9-b360-4d0e6021caf7\") " pod="openshift-marketplace/redhat-marketplace-8nfm5" Nov 24 07:07:58 crc kubenswrapper[4809]: I1124 07:07:58.603360 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e191709-50db-49f9-b360-4d0e6021caf7-utilities\") pod \"redhat-marketplace-8nfm5\" (UID: \"4e191709-50db-49f9-b360-4d0e6021caf7\") " pod="openshift-marketplace/redhat-marketplace-8nfm5" Nov 24 07:07:58 crc kubenswrapper[4809]: I1124 07:07:58.603525 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svkdg\" (UniqueName: \"kubernetes.io/projected/4e191709-50db-49f9-b360-4d0e6021caf7-kube-api-access-svkdg\") pod \"redhat-marketplace-8nfm5\" (UID: \"4e191709-50db-49f9-b360-4d0e6021caf7\") " pod="openshift-marketplace/redhat-marketplace-8nfm5" Nov 24 07:07:58 crc kubenswrapper[4809]: I1124 07:07:58.704645 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e191709-50db-49f9-b360-4d0e6021caf7-catalog-content\") pod \"redhat-marketplace-8nfm5\" (UID: \"4e191709-50db-49f9-b360-4d0e6021caf7\") " pod="openshift-marketplace/redhat-marketplace-8nfm5" Nov 24 07:07:58 crc kubenswrapper[4809]: I1124 07:07:58.704938 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e191709-50db-49f9-b360-4d0e6021caf7-utilities\") pod \"redhat-marketplace-8nfm5\" (UID: \"4e191709-50db-49f9-b360-4d0e6021caf7\") " pod="openshift-marketplace/redhat-marketplace-8nfm5" Nov 24 07:07:58 crc kubenswrapper[4809]: I1124 07:07:58.705003 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svkdg\" (UniqueName: \"kubernetes.io/projected/4e191709-50db-49f9-b360-4d0e6021caf7-kube-api-access-svkdg\") pod \"redhat-marketplace-8nfm5\" (UID: \"4e191709-50db-49f9-b360-4d0e6021caf7\") " pod="openshift-marketplace/redhat-marketplace-8nfm5" Nov 24 07:07:58 crc kubenswrapper[4809]: I1124 07:07:58.705322 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e191709-50db-49f9-b360-4d0e6021caf7-utilities\") pod \"redhat-marketplace-8nfm5\" (UID: \"4e191709-50db-49f9-b360-4d0e6021caf7\") " pod="openshift-marketplace/redhat-marketplace-8nfm5" Nov 24 07:07:58 crc kubenswrapper[4809]: I1124 07:07:58.705486 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e191709-50db-49f9-b360-4d0e6021caf7-catalog-content\") pod \"redhat-marketplace-8nfm5\" (UID: \"4e191709-50db-49f9-b360-4d0e6021caf7\") " pod="openshift-marketplace/redhat-marketplace-8nfm5" Nov 24 07:07:58 crc kubenswrapper[4809]: I1124 07:07:58.741494 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svkdg\" (UniqueName: \"kubernetes.io/projected/4e191709-50db-49f9-b360-4d0e6021caf7-kube-api-access-svkdg\") pod \"redhat-marketplace-8nfm5\" (UID: \"4e191709-50db-49f9-b360-4d0e6021caf7\") " pod="openshift-marketplace/redhat-marketplace-8nfm5" Nov 24 07:07:58 crc kubenswrapper[4809]: I1124 07:07:58.787686 4809 generic.go:334] "Generic (PLEG): container finished" podID="80dfda6c-ae33-4dbd-9750-67079eeacffc" containerID="019669e032c10f26fb695c2cf5d1fb17bc9ad421a4d1392b67d895f60f81c319" exitCode=0 Nov 24 07:07:58 crc kubenswrapper[4809]: I1124 07:07:58.787745 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dp25g" event={"ID":"80dfda6c-ae33-4dbd-9750-67079eeacffc","Type":"ContainerDied","Data":"019669e032c10f26fb695c2cf5d1fb17bc9ad421a4d1392b67d895f60f81c319"} Nov 24 07:07:58 crc kubenswrapper[4809]: I1124 07:07:58.788036 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dp25g" event={"ID":"80dfda6c-ae33-4dbd-9750-67079eeacffc","Type":"ContainerStarted","Data":"86bfafad7245508fdecd8334dc6e09e063c4755e31e54985a5fd02659539de0f"} Nov 24 07:07:58 crc kubenswrapper[4809]: I1124 07:07:58.904933 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8nfm5" Nov 24 07:07:59 crc kubenswrapper[4809]: I1124 07:07:59.293786 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8nfm5"] Nov 24 07:07:59 crc kubenswrapper[4809]: W1124 07:07:59.298958 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e191709_50db_49f9_b360_4d0e6021caf7.slice/crio-2127cd0bd8e704b3fed95b6b05a5f21659c99c20a1958d7db925d4b70b6ffd7a WatchSource:0}: Error finding container 2127cd0bd8e704b3fed95b6b05a5f21659c99c20a1958d7db925d4b70b6ffd7a: Status 404 returned error can't find the container with id 2127cd0bd8e704b3fed95b6b05a5f21659c99c20a1958d7db925d4b70b6ffd7a Nov 24 07:07:59 crc kubenswrapper[4809]: I1124 07:07:59.796296 4809 generic.go:334] "Generic (PLEG): container finished" podID="4e191709-50db-49f9-b360-4d0e6021caf7" containerID="a325461f1678cf64695e6e8ac7986eb9bbd67200cd593e39b8512478e84d7acb" exitCode=0 Nov 24 07:07:59 crc kubenswrapper[4809]: I1124 07:07:59.796387 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nfm5" event={"ID":"4e191709-50db-49f9-b360-4d0e6021caf7","Type":"ContainerDied","Data":"a325461f1678cf64695e6e8ac7986eb9bbd67200cd593e39b8512478e84d7acb"} Nov 24 07:07:59 crc kubenswrapper[4809]: I1124 07:07:59.796424 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nfm5" event={"ID":"4e191709-50db-49f9-b360-4d0e6021caf7","Type":"ContainerStarted","Data":"2127cd0bd8e704b3fed95b6b05a5f21659c99c20a1958d7db925d4b70b6ffd7a"} Nov 24 07:07:59 crc kubenswrapper[4809]: I1124 07:07:59.800170 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dp25g" event={"ID":"80dfda6c-ae33-4dbd-9750-67079eeacffc","Type":"ContainerStarted","Data":"d4653ea4dd3999ce95f3676551840687ab4008e9aae80e24520d062284d019e9"} Nov 24 07:08:00 crc kubenswrapper[4809]: I1124 07:08:00.808682 4809 generic.go:334] "Generic (PLEG): container finished" podID="4e191709-50db-49f9-b360-4d0e6021caf7" containerID="58debb25447c4674447f51586880604ea37a5e7e6828a102f0e1edb8d0c9b0ff" exitCode=0 Nov 24 07:08:00 crc kubenswrapper[4809]: I1124 07:08:00.808803 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nfm5" event={"ID":"4e191709-50db-49f9-b360-4d0e6021caf7","Type":"ContainerDied","Data":"58debb25447c4674447f51586880604ea37a5e7e6828a102f0e1edb8d0c9b0ff"} Nov 24 07:08:00 crc kubenswrapper[4809]: I1124 07:08:00.811848 4809 generic.go:334] "Generic (PLEG): container finished" podID="80dfda6c-ae33-4dbd-9750-67079eeacffc" containerID="d4653ea4dd3999ce95f3676551840687ab4008e9aae80e24520d062284d019e9" exitCode=0 Nov 24 07:08:00 crc kubenswrapper[4809]: I1124 07:08:00.811878 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dp25g" event={"ID":"80dfda6c-ae33-4dbd-9750-67079eeacffc","Type":"ContainerDied","Data":"d4653ea4dd3999ce95f3676551840687ab4008e9aae80e24520d062284d019e9"} Nov 24 07:08:01 crc kubenswrapper[4809]: I1124 07:08:01.821075 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dp25g" event={"ID":"80dfda6c-ae33-4dbd-9750-67079eeacffc","Type":"ContainerStarted","Data":"e76ea9e47d6554c7208ec3ec9c105d059b45d9e20bb3fb375410703e9e11cb6a"} Nov 24 07:08:01 crc kubenswrapper[4809]: I1124 07:08:01.823538 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nfm5" event={"ID":"4e191709-50db-49f9-b360-4d0e6021caf7","Type":"ContainerStarted","Data":"b14950cb40332cb830346aecae53e5490b5f1811ca3e0fdd70fa4924c770adf4"} Nov 24 07:08:01 crc kubenswrapper[4809]: I1124 07:08:01.842368 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dp25g" podStartSLOduration=2.3444115549999998 podStartE2EDuration="4.84235024s" podCreationTimestamp="2025-11-24 07:07:57 +0000 UTC" firstStartedPulling="2025-11-24 07:07:58.789016728 +0000 UTC m=+778.689608433" lastFinishedPulling="2025-11-24 07:08:01.286955373 +0000 UTC m=+781.187547118" observedRunningTime="2025-11-24 07:08:01.841015767 +0000 UTC m=+781.741607472" watchObservedRunningTime="2025-11-24 07:08:01.84235024 +0000 UTC m=+781.742941955" Nov 24 07:08:01 crc kubenswrapper[4809]: I1124 07:08:01.863009 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8nfm5" podStartSLOduration=2.4773965479999998 podStartE2EDuration="3.862988811s" podCreationTimestamp="2025-11-24 07:07:58 +0000 UTC" firstStartedPulling="2025-11-24 07:07:59.798206499 +0000 UTC m=+779.698798204" lastFinishedPulling="2025-11-24 07:08:01.183798722 +0000 UTC m=+781.084390467" observedRunningTime="2025-11-24 07:08:01.861162635 +0000 UTC m=+781.761754340" watchObservedRunningTime="2025-11-24 07:08:01.862988811 +0000 UTC m=+781.763580516" Nov 24 07:08:03 crc kubenswrapper[4809]: I1124 07:08:03.917726 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-mhj4l" Nov 24 07:08:03 crc kubenswrapper[4809]: I1124 07:08:03.918227 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-mhj4l" Nov 24 07:08:03 crc kubenswrapper[4809]: I1124 07:08:03.956855 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-mhj4l" Nov 24 07:08:04 crc kubenswrapper[4809]: I1124 07:08:04.873423 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-mhj4l" Nov 24 07:08:05 crc kubenswrapper[4809]: I1124 07:08:05.728108 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mn5w6"] Nov 24 07:08:05 crc kubenswrapper[4809]: I1124 07:08:05.731351 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mn5w6" Nov 24 07:08:05 crc kubenswrapper[4809]: I1124 07:08:05.749434 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mn5w6"] Nov 24 07:08:05 crc kubenswrapper[4809]: I1124 07:08:05.849106 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d60f4600-cb92-4b50-a958-caba8079a0fb-utilities\") pod \"certified-operators-mn5w6\" (UID: \"d60f4600-cb92-4b50-a958-caba8079a0fb\") " pod="openshift-marketplace/certified-operators-mn5w6" Nov 24 07:08:05 crc kubenswrapper[4809]: I1124 07:08:05.849209 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lbt9\" (UniqueName: \"kubernetes.io/projected/d60f4600-cb92-4b50-a958-caba8079a0fb-kube-api-access-4lbt9\") pod \"certified-operators-mn5w6\" (UID: \"d60f4600-cb92-4b50-a958-caba8079a0fb\") " pod="openshift-marketplace/certified-operators-mn5w6" Nov 24 07:08:05 crc kubenswrapper[4809]: I1124 07:08:05.849607 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d60f4600-cb92-4b50-a958-caba8079a0fb-catalog-content\") pod \"certified-operators-mn5w6\" (UID: \"d60f4600-cb92-4b50-a958-caba8079a0fb\") " pod="openshift-marketplace/certified-operators-mn5w6" Nov 24 07:08:05 crc kubenswrapper[4809]: I1124 07:08:05.951442 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d60f4600-cb92-4b50-a958-caba8079a0fb-utilities\") pod \"certified-operators-mn5w6\" (UID: \"d60f4600-cb92-4b50-a958-caba8079a0fb\") " pod="openshift-marketplace/certified-operators-mn5w6" Nov 24 07:08:05 crc kubenswrapper[4809]: I1124 07:08:05.951569 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lbt9\" (UniqueName: \"kubernetes.io/projected/d60f4600-cb92-4b50-a958-caba8079a0fb-kube-api-access-4lbt9\") pod \"certified-operators-mn5w6\" (UID: \"d60f4600-cb92-4b50-a958-caba8079a0fb\") " pod="openshift-marketplace/certified-operators-mn5w6" Nov 24 07:08:05 crc kubenswrapper[4809]: I1124 07:08:05.951738 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d60f4600-cb92-4b50-a958-caba8079a0fb-catalog-content\") pod \"certified-operators-mn5w6\" (UID: \"d60f4600-cb92-4b50-a958-caba8079a0fb\") " pod="openshift-marketplace/certified-operators-mn5w6" Nov 24 07:08:05 crc kubenswrapper[4809]: I1124 07:08:05.951957 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d60f4600-cb92-4b50-a958-caba8079a0fb-utilities\") pod \"certified-operators-mn5w6\" (UID: \"d60f4600-cb92-4b50-a958-caba8079a0fb\") " pod="openshift-marketplace/certified-operators-mn5w6" Nov 24 07:08:05 crc kubenswrapper[4809]: I1124 07:08:05.952221 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d60f4600-cb92-4b50-a958-caba8079a0fb-catalog-content\") pod \"certified-operators-mn5w6\" (UID: \"d60f4600-cb92-4b50-a958-caba8079a0fb\") " pod="openshift-marketplace/certified-operators-mn5w6" Nov 24 07:08:05 crc kubenswrapper[4809]: I1124 07:08:05.975434 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lbt9\" (UniqueName: \"kubernetes.io/projected/d60f4600-cb92-4b50-a958-caba8079a0fb-kube-api-access-4lbt9\") pod \"certified-operators-mn5w6\" (UID: \"d60f4600-cb92-4b50-a958-caba8079a0fb\") " pod="openshift-marketplace/certified-operators-mn5w6" Nov 24 07:08:06 crc kubenswrapper[4809]: I1124 07:08:06.084002 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mn5w6" Nov 24 07:08:06 crc kubenswrapper[4809]: I1124 07:08:06.547877 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mn5w6"] Nov 24 07:08:06 crc kubenswrapper[4809]: I1124 07:08:06.858761 4809 generic.go:334] "Generic (PLEG): container finished" podID="d60f4600-cb92-4b50-a958-caba8079a0fb" containerID="194ce5401fedba62813c939d956179a7ae74745293b273230a72f946eb2fd624" exitCode=0 Nov 24 07:08:06 crc kubenswrapper[4809]: I1124 07:08:06.858832 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mn5w6" event={"ID":"d60f4600-cb92-4b50-a958-caba8079a0fb","Type":"ContainerDied","Data":"194ce5401fedba62813c939d956179a7ae74745293b273230a72f946eb2fd624"} Nov 24 07:08:06 crc kubenswrapper[4809]: I1124 07:08:06.859765 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mn5w6" event={"ID":"d60f4600-cb92-4b50-a958-caba8079a0fb","Type":"ContainerStarted","Data":"44b5a2c81bc458e88fd3dcc346fd81ae2cfa1fa888ad992fc963aced6f4e9f11"} Nov 24 07:08:07 crc kubenswrapper[4809]: I1124 07:08:07.873904 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mn5w6" event={"ID":"d60f4600-cb92-4b50-a958-caba8079a0fb","Type":"ContainerStarted","Data":"ab769e086a07620a47d65fb0394e990ce13ea4f06c5c584f47fbd8195ead7306"} Nov 24 07:08:08 crc kubenswrapper[4809]: I1124 07:08:08.050416 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dp25g" Nov 24 07:08:08 crc kubenswrapper[4809]: I1124 07:08:08.050521 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dp25g" Nov 24 07:08:08 crc kubenswrapper[4809]: I1124 07:08:08.129083 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dp25g" Nov 24 07:08:08 crc kubenswrapper[4809]: I1124 07:08:08.885935 4809 generic.go:334] "Generic (PLEG): container finished" podID="d60f4600-cb92-4b50-a958-caba8079a0fb" containerID="ab769e086a07620a47d65fb0394e990ce13ea4f06c5c584f47fbd8195ead7306" exitCode=0 Nov 24 07:08:08 crc kubenswrapper[4809]: I1124 07:08:08.886050 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mn5w6" event={"ID":"d60f4600-cb92-4b50-a958-caba8079a0fb","Type":"ContainerDied","Data":"ab769e086a07620a47d65fb0394e990ce13ea4f06c5c584f47fbd8195ead7306"} Nov 24 07:08:08 crc kubenswrapper[4809]: I1124 07:08:08.905359 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8nfm5" Nov 24 07:08:08 crc kubenswrapper[4809]: I1124 07:08:08.905437 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8nfm5" Nov 24 07:08:09 crc kubenswrapper[4809]: I1124 07:08:09.007533 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dp25g" Nov 24 07:08:09 crc kubenswrapper[4809]: I1124 07:08:09.008475 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8nfm5" Nov 24 07:08:09 crc kubenswrapper[4809]: I1124 07:08:09.897538 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mn5w6" event={"ID":"d60f4600-cb92-4b50-a958-caba8079a0fb","Type":"ContainerStarted","Data":"2afd482ef9ba62f193fa2c16f796f33ba0e88cb99591ee995fb5d690f343a440"} Nov 24 07:08:09 crc kubenswrapper[4809]: I1124 07:08:09.927751 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mn5w6" podStartSLOduration=2.475322401 podStartE2EDuration="4.927737929s" podCreationTimestamp="2025-11-24 07:08:05 +0000 UTC" firstStartedPulling="2025-11-24 07:08:06.864055494 +0000 UTC m=+786.764647209" lastFinishedPulling="2025-11-24 07:08:09.316471022 +0000 UTC m=+789.217062737" observedRunningTime="2025-11-24 07:08:09.925948163 +0000 UTC m=+789.826539868" watchObservedRunningTime="2025-11-24 07:08:09.927737929 +0000 UTC m=+789.828329634" Nov 24 07:08:09 crc kubenswrapper[4809]: I1124 07:08:09.964567 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8nfm5" Nov 24 07:08:11 crc kubenswrapper[4809]: I1124 07:08:11.965728 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b"] Nov 24 07:08:11 crc kubenswrapper[4809]: I1124 07:08:11.968040 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b" Nov 24 07:08:11 crc kubenswrapper[4809]: I1124 07:08:11.970998 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-nv54p" Nov 24 07:08:11 crc kubenswrapper[4809]: I1124 07:08:11.983440 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b"] Nov 24 07:08:12 crc kubenswrapper[4809]: I1124 07:08:12.140829 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ed47ae85-6bd2-4e58-acee-30c9022b9170-bundle\") pod \"33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b\" (UID: \"ed47ae85-6bd2-4e58-acee-30c9022b9170\") " pod="openstack-operators/33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b" Nov 24 07:08:12 crc kubenswrapper[4809]: I1124 07:08:12.140921 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk4qb\" (UniqueName: \"kubernetes.io/projected/ed47ae85-6bd2-4e58-acee-30c9022b9170-kube-api-access-lk4qb\") pod \"33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b\" (UID: \"ed47ae85-6bd2-4e58-acee-30c9022b9170\") " pod="openstack-operators/33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b" Nov 24 07:08:12 crc kubenswrapper[4809]: I1124 07:08:12.140948 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ed47ae85-6bd2-4e58-acee-30c9022b9170-util\") pod \"33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b\" (UID: \"ed47ae85-6bd2-4e58-acee-30c9022b9170\") " pod="openstack-operators/33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b" Nov 24 07:08:12 crc kubenswrapper[4809]: I1124 07:08:12.241832 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ed47ae85-6bd2-4e58-acee-30c9022b9170-bundle\") pod \"33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b\" (UID: \"ed47ae85-6bd2-4e58-acee-30c9022b9170\") " pod="openstack-operators/33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b" Nov 24 07:08:12 crc kubenswrapper[4809]: I1124 07:08:12.242249 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk4qb\" (UniqueName: \"kubernetes.io/projected/ed47ae85-6bd2-4e58-acee-30c9022b9170-kube-api-access-lk4qb\") pod \"33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b\" (UID: \"ed47ae85-6bd2-4e58-acee-30c9022b9170\") " pod="openstack-operators/33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b" Nov 24 07:08:12 crc kubenswrapper[4809]: I1124 07:08:12.242333 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ed47ae85-6bd2-4e58-acee-30c9022b9170-util\") pod \"33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b\" (UID: \"ed47ae85-6bd2-4e58-acee-30c9022b9170\") " pod="openstack-operators/33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b" Nov 24 07:08:12 crc kubenswrapper[4809]: I1124 07:08:12.242682 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ed47ae85-6bd2-4e58-acee-30c9022b9170-bundle\") pod \"33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b\" (UID: \"ed47ae85-6bd2-4e58-acee-30c9022b9170\") " pod="openstack-operators/33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b" Nov 24 07:08:12 crc kubenswrapper[4809]: I1124 07:08:12.243089 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ed47ae85-6bd2-4e58-acee-30c9022b9170-util\") pod \"33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b\" (UID: \"ed47ae85-6bd2-4e58-acee-30c9022b9170\") " pod="openstack-operators/33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b" Nov 24 07:08:12 crc kubenswrapper[4809]: I1124 07:08:12.263352 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk4qb\" (UniqueName: \"kubernetes.io/projected/ed47ae85-6bd2-4e58-acee-30c9022b9170-kube-api-access-lk4qb\") pod \"33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b\" (UID: \"ed47ae85-6bd2-4e58-acee-30c9022b9170\") " pod="openstack-operators/33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b" Nov 24 07:08:12 crc kubenswrapper[4809]: I1124 07:08:12.286311 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b" Nov 24 07:08:12 crc kubenswrapper[4809]: I1124 07:08:12.511949 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dp25g"] Nov 24 07:08:12 crc kubenswrapper[4809]: I1124 07:08:12.512587 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dp25g" podUID="80dfda6c-ae33-4dbd-9750-67079eeacffc" containerName="registry-server" containerID="cri-o://e76ea9e47d6554c7208ec3ec9c105d059b45d9e20bb3fb375410703e9e11cb6a" gracePeriod=2 Nov 24 07:08:12 crc kubenswrapper[4809]: I1124 07:08:12.782618 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b"] Nov 24 07:08:12 crc kubenswrapper[4809]: W1124 07:08:12.795161 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded47ae85_6bd2_4e58_acee_30c9022b9170.slice/crio-ff60b9021d3741bcf16056a09e2a04eb6c956914ac7deec20fc249378c3357d5 WatchSource:0}: Error finding container ff60b9021d3741bcf16056a09e2a04eb6c956914ac7deec20fc249378c3357d5: Status 404 returned error can't find the container with id ff60b9021d3741bcf16056a09e2a04eb6c956914ac7deec20fc249378c3357d5 Nov 24 07:08:12 crc kubenswrapper[4809]: I1124 07:08:12.893061 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dp25g" Nov 24 07:08:12 crc kubenswrapper[4809]: I1124 07:08:12.921864 4809 generic.go:334] "Generic (PLEG): container finished" podID="80dfda6c-ae33-4dbd-9750-67079eeacffc" containerID="e76ea9e47d6554c7208ec3ec9c105d059b45d9e20bb3fb375410703e9e11cb6a" exitCode=0 Nov 24 07:08:12 crc kubenswrapper[4809]: I1124 07:08:12.921930 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dp25g" event={"ID":"80dfda6c-ae33-4dbd-9750-67079eeacffc","Type":"ContainerDied","Data":"e76ea9e47d6554c7208ec3ec9c105d059b45d9e20bb3fb375410703e9e11cb6a"} Nov 24 07:08:12 crc kubenswrapper[4809]: I1124 07:08:12.921980 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dp25g" event={"ID":"80dfda6c-ae33-4dbd-9750-67079eeacffc","Type":"ContainerDied","Data":"86bfafad7245508fdecd8334dc6e09e063c4755e31e54985a5fd02659539de0f"} Nov 24 07:08:12 crc kubenswrapper[4809]: I1124 07:08:12.921999 4809 scope.go:117] "RemoveContainer" containerID="e76ea9e47d6554c7208ec3ec9c105d059b45d9e20bb3fb375410703e9e11cb6a" Nov 24 07:08:12 crc kubenswrapper[4809]: I1124 07:08:12.922107 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dp25g" Nov 24 07:08:12 crc kubenswrapper[4809]: I1124 07:08:12.931752 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b" event={"ID":"ed47ae85-6bd2-4e58-acee-30c9022b9170","Type":"ContainerStarted","Data":"ff60b9021d3741bcf16056a09e2a04eb6c956914ac7deec20fc249378c3357d5"} Nov 24 07:08:12 crc kubenswrapper[4809]: I1124 07:08:12.949498 4809 scope.go:117] "RemoveContainer" containerID="d4653ea4dd3999ce95f3676551840687ab4008e9aae80e24520d062284d019e9" Nov 24 07:08:12 crc kubenswrapper[4809]: I1124 07:08:12.972719 4809 scope.go:117] "RemoveContainer" containerID="019669e032c10f26fb695c2cf5d1fb17bc9ad421a4d1392b67d895f60f81c319" Nov 24 07:08:13 crc kubenswrapper[4809]: I1124 07:08:13.006678 4809 scope.go:117] "RemoveContainer" containerID="e76ea9e47d6554c7208ec3ec9c105d059b45d9e20bb3fb375410703e9e11cb6a" Nov 24 07:08:13 crc kubenswrapper[4809]: E1124 07:08:13.007118 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e76ea9e47d6554c7208ec3ec9c105d059b45d9e20bb3fb375410703e9e11cb6a\": container with ID starting with e76ea9e47d6554c7208ec3ec9c105d059b45d9e20bb3fb375410703e9e11cb6a not found: ID does not exist" containerID="e76ea9e47d6554c7208ec3ec9c105d059b45d9e20bb3fb375410703e9e11cb6a" Nov 24 07:08:13 crc kubenswrapper[4809]: I1124 07:08:13.007162 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e76ea9e47d6554c7208ec3ec9c105d059b45d9e20bb3fb375410703e9e11cb6a"} err="failed to get container status \"e76ea9e47d6554c7208ec3ec9c105d059b45d9e20bb3fb375410703e9e11cb6a\": rpc error: code = NotFound desc = could not find container \"e76ea9e47d6554c7208ec3ec9c105d059b45d9e20bb3fb375410703e9e11cb6a\": container with ID starting with e76ea9e47d6554c7208ec3ec9c105d059b45d9e20bb3fb375410703e9e11cb6a not found: ID does not exist" Nov 24 07:08:13 crc kubenswrapper[4809]: I1124 07:08:13.007196 4809 scope.go:117] "RemoveContainer" containerID="d4653ea4dd3999ce95f3676551840687ab4008e9aae80e24520d062284d019e9" Nov 24 07:08:13 crc kubenswrapper[4809]: E1124 07:08:13.010682 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4653ea4dd3999ce95f3676551840687ab4008e9aae80e24520d062284d019e9\": container with ID starting with d4653ea4dd3999ce95f3676551840687ab4008e9aae80e24520d062284d019e9 not found: ID does not exist" containerID="d4653ea4dd3999ce95f3676551840687ab4008e9aae80e24520d062284d019e9" Nov 24 07:08:13 crc kubenswrapper[4809]: I1124 07:08:13.010719 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4653ea4dd3999ce95f3676551840687ab4008e9aae80e24520d062284d019e9"} err="failed to get container status \"d4653ea4dd3999ce95f3676551840687ab4008e9aae80e24520d062284d019e9\": rpc error: code = NotFound desc = could not find container \"d4653ea4dd3999ce95f3676551840687ab4008e9aae80e24520d062284d019e9\": container with ID starting with d4653ea4dd3999ce95f3676551840687ab4008e9aae80e24520d062284d019e9 not found: ID does not exist" Nov 24 07:08:13 crc kubenswrapper[4809]: I1124 07:08:13.010758 4809 scope.go:117] "RemoveContainer" containerID="019669e032c10f26fb695c2cf5d1fb17bc9ad421a4d1392b67d895f60f81c319" Nov 24 07:08:13 crc kubenswrapper[4809]: E1124 07:08:13.011290 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"019669e032c10f26fb695c2cf5d1fb17bc9ad421a4d1392b67d895f60f81c319\": container with ID starting with 019669e032c10f26fb695c2cf5d1fb17bc9ad421a4d1392b67d895f60f81c319 not found: ID does not exist" containerID="019669e032c10f26fb695c2cf5d1fb17bc9ad421a4d1392b67d895f60f81c319" Nov 24 07:08:13 crc kubenswrapper[4809]: I1124 07:08:13.011315 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"019669e032c10f26fb695c2cf5d1fb17bc9ad421a4d1392b67d895f60f81c319"} err="failed to get container status \"019669e032c10f26fb695c2cf5d1fb17bc9ad421a4d1392b67d895f60f81c319\": rpc error: code = NotFound desc = could not find container \"019669e032c10f26fb695c2cf5d1fb17bc9ad421a4d1392b67d895f60f81c319\": container with ID starting with 019669e032c10f26fb695c2cf5d1fb17bc9ad421a4d1392b67d895f60f81c319 not found: ID does not exist" Nov 24 07:08:13 crc kubenswrapper[4809]: I1124 07:08:13.053560 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57vlq\" (UniqueName: \"kubernetes.io/projected/80dfda6c-ae33-4dbd-9750-67079eeacffc-kube-api-access-57vlq\") pod \"80dfda6c-ae33-4dbd-9750-67079eeacffc\" (UID: \"80dfda6c-ae33-4dbd-9750-67079eeacffc\") " Nov 24 07:08:13 crc kubenswrapper[4809]: I1124 07:08:13.053658 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80dfda6c-ae33-4dbd-9750-67079eeacffc-catalog-content\") pod \"80dfda6c-ae33-4dbd-9750-67079eeacffc\" (UID: \"80dfda6c-ae33-4dbd-9750-67079eeacffc\") " Nov 24 07:08:13 crc kubenswrapper[4809]: I1124 07:08:13.053751 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80dfda6c-ae33-4dbd-9750-67079eeacffc-utilities\") pod \"80dfda6c-ae33-4dbd-9750-67079eeacffc\" (UID: \"80dfda6c-ae33-4dbd-9750-67079eeacffc\") " Nov 24 07:08:13 crc kubenswrapper[4809]: I1124 07:08:13.055743 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80dfda6c-ae33-4dbd-9750-67079eeacffc-utilities" (OuterVolumeSpecName: "utilities") pod "80dfda6c-ae33-4dbd-9750-67079eeacffc" (UID: "80dfda6c-ae33-4dbd-9750-67079eeacffc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:08:13 crc kubenswrapper[4809]: I1124 07:08:13.059051 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80dfda6c-ae33-4dbd-9750-67079eeacffc-kube-api-access-57vlq" (OuterVolumeSpecName: "kube-api-access-57vlq") pod "80dfda6c-ae33-4dbd-9750-67079eeacffc" (UID: "80dfda6c-ae33-4dbd-9750-67079eeacffc"). InnerVolumeSpecName "kube-api-access-57vlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:08:13 crc kubenswrapper[4809]: I1124 07:08:13.102524 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80dfda6c-ae33-4dbd-9750-67079eeacffc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "80dfda6c-ae33-4dbd-9750-67079eeacffc" (UID: "80dfda6c-ae33-4dbd-9750-67079eeacffc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:08:13 crc kubenswrapper[4809]: I1124 07:08:13.155681 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80dfda6c-ae33-4dbd-9750-67079eeacffc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:13 crc kubenswrapper[4809]: I1124 07:08:13.155718 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80dfda6c-ae33-4dbd-9750-67079eeacffc-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:13 crc kubenswrapper[4809]: I1124 07:08:13.155728 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57vlq\" (UniqueName: \"kubernetes.io/projected/80dfda6c-ae33-4dbd-9750-67079eeacffc-kube-api-access-57vlq\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:13 crc kubenswrapper[4809]: I1124 07:08:13.249065 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dp25g"] Nov 24 07:08:13 crc kubenswrapper[4809]: I1124 07:08:13.252181 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dp25g"] Nov 24 07:08:13 crc kubenswrapper[4809]: I1124 07:08:13.929724 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8nfm5"] Nov 24 07:08:13 crc kubenswrapper[4809]: I1124 07:08:13.930537 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8nfm5" podUID="4e191709-50db-49f9-b360-4d0e6021caf7" containerName="registry-server" containerID="cri-o://b14950cb40332cb830346aecae53e5490b5f1811ca3e0fdd70fa4924c770adf4" gracePeriod=2 Nov 24 07:08:13 crc kubenswrapper[4809]: I1124 07:08:13.942247 4809 generic.go:334] "Generic (PLEG): container finished" podID="ed47ae85-6bd2-4e58-acee-30c9022b9170" containerID="2cce0d9732879d0cbf0f843ae209cbb16401795fa74e0ceb68c3741c8c71e78d" exitCode=0 Nov 24 07:08:13 crc kubenswrapper[4809]: I1124 07:08:13.942356 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b" event={"ID":"ed47ae85-6bd2-4e58-acee-30c9022b9170","Type":"ContainerDied","Data":"2cce0d9732879d0cbf0f843ae209cbb16401795fa74e0ceb68c3741c8c71e78d"} Nov 24 07:08:14 crc kubenswrapper[4809]: I1124 07:08:14.486570 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8nfm5" Nov 24 07:08:14 crc kubenswrapper[4809]: I1124 07:08:14.576631 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e191709-50db-49f9-b360-4d0e6021caf7-utilities\") pod \"4e191709-50db-49f9-b360-4d0e6021caf7\" (UID: \"4e191709-50db-49f9-b360-4d0e6021caf7\") " Nov 24 07:08:14 crc kubenswrapper[4809]: I1124 07:08:14.576776 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e191709-50db-49f9-b360-4d0e6021caf7-catalog-content\") pod \"4e191709-50db-49f9-b360-4d0e6021caf7\" (UID: \"4e191709-50db-49f9-b360-4d0e6021caf7\") " Nov 24 07:08:14 crc kubenswrapper[4809]: I1124 07:08:14.576810 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svkdg\" (UniqueName: \"kubernetes.io/projected/4e191709-50db-49f9-b360-4d0e6021caf7-kube-api-access-svkdg\") pod \"4e191709-50db-49f9-b360-4d0e6021caf7\" (UID: \"4e191709-50db-49f9-b360-4d0e6021caf7\") " Nov 24 07:08:14 crc kubenswrapper[4809]: I1124 07:08:14.577747 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e191709-50db-49f9-b360-4d0e6021caf7-utilities" (OuterVolumeSpecName: "utilities") pod "4e191709-50db-49f9-b360-4d0e6021caf7" (UID: "4e191709-50db-49f9-b360-4d0e6021caf7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:08:14 crc kubenswrapper[4809]: I1124 07:08:14.583661 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e191709-50db-49f9-b360-4d0e6021caf7-kube-api-access-svkdg" (OuterVolumeSpecName: "kube-api-access-svkdg") pod "4e191709-50db-49f9-b360-4d0e6021caf7" (UID: "4e191709-50db-49f9-b360-4d0e6021caf7"). InnerVolumeSpecName "kube-api-access-svkdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:08:14 crc kubenswrapper[4809]: I1124 07:08:14.591719 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e191709-50db-49f9-b360-4d0e6021caf7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4e191709-50db-49f9-b360-4d0e6021caf7" (UID: "4e191709-50db-49f9-b360-4d0e6021caf7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:08:14 crc kubenswrapper[4809]: I1124 07:08:14.677902 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e191709-50db-49f9-b360-4d0e6021caf7-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:14 crc kubenswrapper[4809]: I1124 07:08:14.677949 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e191709-50db-49f9-b360-4d0e6021caf7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:14 crc kubenswrapper[4809]: I1124 07:08:14.678006 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svkdg\" (UniqueName: \"kubernetes.io/projected/4e191709-50db-49f9-b360-4d0e6021caf7-kube-api-access-svkdg\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:14 crc kubenswrapper[4809]: I1124 07:08:14.900068 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80dfda6c-ae33-4dbd-9750-67079eeacffc" path="/var/lib/kubelet/pods/80dfda6c-ae33-4dbd-9750-67079eeacffc/volumes" Nov 24 07:08:14 crc kubenswrapper[4809]: I1124 07:08:14.954393 4809 generic.go:334] "Generic (PLEG): container finished" podID="ed47ae85-6bd2-4e58-acee-30c9022b9170" containerID="8bf49fc8e21781ab50a69222bde9ddea60413ed7d6c222d275ce97275e219da5" exitCode=0 Nov 24 07:08:14 crc kubenswrapper[4809]: I1124 07:08:14.954469 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b" event={"ID":"ed47ae85-6bd2-4e58-acee-30c9022b9170","Type":"ContainerDied","Data":"8bf49fc8e21781ab50a69222bde9ddea60413ed7d6c222d275ce97275e219da5"} Nov 24 07:08:14 crc kubenswrapper[4809]: I1124 07:08:14.961266 4809 generic.go:334] "Generic (PLEG): container finished" podID="4e191709-50db-49f9-b360-4d0e6021caf7" containerID="b14950cb40332cb830346aecae53e5490b5f1811ca3e0fdd70fa4924c770adf4" exitCode=0 Nov 24 07:08:14 crc kubenswrapper[4809]: I1124 07:08:14.961325 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nfm5" event={"ID":"4e191709-50db-49f9-b360-4d0e6021caf7","Type":"ContainerDied","Data":"b14950cb40332cb830346aecae53e5490b5f1811ca3e0fdd70fa4924c770adf4"} Nov 24 07:08:14 crc kubenswrapper[4809]: I1124 07:08:14.961366 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8nfm5" Nov 24 07:08:14 crc kubenswrapper[4809]: I1124 07:08:14.961401 4809 scope.go:117] "RemoveContainer" containerID="b14950cb40332cb830346aecae53e5490b5f1811ca3e0fdd70fa4924c770adf4" Nov 24 07:08:14 crc kubenswrapper[4809]: I1124 07:08:14.961375 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nfm5" event={"ID":"4e191709-50db-49f9-b360-4d0e6021caf7","Type":"ContainerDied","Data":"2127cd0bd8e704b3fed95b6b05a5f21659c99c20a1958d7db925d4b70b6ffd7a"} Nov 24 07:08:15 crc kubenswrapper[4809]: I1124 07:08:15.002547 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8nfm5"] Nov 24 07:08:15 crc kubenswrapper[4809]: I1124 07:08:15.003188 4809 scope.go:117] "RemoveContainer" containerID="58debb25447c4674447f51586880604ea37a5e7e6828a102f0e1edb8d0c9b0ff" Nov 24 07:08:15 crc kubenswrapper[4809]: I1124 07:08:15.005729 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8nfm5"] Nov 24 07:08:15 crc kubenswrapper[4809]: I1124 07:08:15.042554 4809 scope.go:117] "RemoveContainer" containerID="a325461f1678cf64695e6e8ac7986eb9bbd67200cd593e39b8512478e84d7acb" Nov 24 07:08:15 crc kubenswrapper[4809]: I1124 07:08:15.092409 4809 scope.go:117] "RemoveContainer" containerID="b14950cb40332cb830346aecae53e5490b5f1811ca3e0fdd70fa4924c770adf4" Nov 24 07:08:15 crc kubenswrapper[4809]: E1124 07:08:15.093041 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b14950cb40332cb830346aecae53e5490b5f1811ca3e0fdd70fa4924c770adf4\": container with ID starting with b14950cb40332cb830346aecae53e5490b5f1811ca3e0fdd70fa4924c770adf4 not found: ID does not exist" containerID="b14950cb40332cb830346aecae53e5490b5f1811ca3e0fdd70fa4924c770adf4" Nov 24 07:08:15 crc kubenswrapper[4809]: I1124 07:08:15.093079 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b14950cb40332cb830346aecae53e5490b5f1811ca3e0fdd70fa4924c770adf4"} err="failed to get container status \"b14950cb40332cb830346aecae53e5490b5f1811ca3e0fdd70fa4924c770adf4\": rpc error: code = NotFound desc = could not find container \"b14950cb40332cb830346aecae53e5490b5f1811ca3e0fdd70fa4924c770adf4\": container with ID starting with b14950cb40332cb830346aecae53e5490b5f1811ca3e0fdd70fa4924c770adf4 not found: ID does not exist" Nov 24 07:08:15 crc kubenswrapper[4809]: I1124 07:08:15.093105 4809 scope.go:117] "RemoveContainer" containerID="58debb25447c4674447f51586880604ea37a5e7e6828a102f0e1edb8d0c9b0ff" Nov 24 07:08:15 crc kubenswrapper[4809]: E1124 07:08:15.093538 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58debb25447c4674447f51586880604ea37a5e7e6828a102f0e1edb8d0c9b0ff\": container with ID starting with 58debb25447c4674447f51586880604ea37a5e7e6828a102f0e1edb8d0c9b0ff not found: ID does not exist" containerID="58debb25447c4674447f51586880604ea37a5e7e6828a102f0e1edb8d0c9b0ff" Nov 24 07:08:15 crc kubenswrapper[4809]: I1124 07:08:15.093570 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58debb25447c4674447f51586880604ea37a5e7e6828a102f0e1edb8d0c9b0ff"} err="failed to get container status \"58debb25447c4674447f51586880604ea37a5e7e6828a102f0e1edb8d0c9b0ff\": rpc error: code = NotFound desc = could not find container \"58debb25447c4674447f51586880604ea37a5e7e6828a102f0e1edb8d0c9b0ff\": container with ID starting with 58debb25447c4674447f51586880604ea37a5e7e6828a102f0e1edb8d0c9b0ff not found: ID does not exist" Nov 24 07:08:15 crc kubenswrapper[4809]: I1124 07:08:15.093608 4809 scope.go:117] "RemoveContainer" containerID="a325461f1678cf64695e6e8ac7986eb9bbd67200cd593e39b8512478e84d7acb" Nov 24 07:08:15 crc kubenswrapper[4809]: E1124 07:08:15.094708 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a325461f1678cf64695e6e8ac7986eb9bbd67200cd593e39b8512478e84d7acb\": container with ID starting with a325461f1678cf64695e6e8ac7986eb9bbd67200cd593e39b8512478e84d7acb not found: ID does not exist" containerID="a325461f1678cf64695e6e8ac7986eb9bbd67200cd593e39b8512478e84d7acb" Nov 24 07:08:15 crc kubenswrapper[4809]: I1124 07:08:15.094737 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a325461f1678cf64695e6e8ac7986eb9bbd67200cd593e39b8512478e84d7acb"} err="failed to get container status \"a325461f1678cf64695e6e8ac7986eb9bbd67200cd593e39b8512478e84d7acb\": rpc error: code = NotFound desc = could not find container \"a325461f1678cf64695e6e8ac7986eb9bbd67200cd593e39b8512478e84d7acb\": container with ID starting with a325461f1678cf64695e6e8ac7986eb9bbd67200cd593e39b8512478e84d7acb not found: ID does not exist" Nov 24 07:08:15 crc kubenswrapper[4809]: I1124 07:08:15.970213 4809 generic.go:334] "Generic (PLEG): container finished" podID="ed47ae85-6bd2-4e58-acee-30c9022b9170" containerID="e0bb131dcce7694c9a63e34597dce47a19e480e2578bc471243f2024f6cd4bca" exitCode=0 Nov 24 07:08:15 crc kubenswrapper[4809]: I1124 07:08:15.970298 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b" event={"ID":"ed47ae85-6bd2-4e58-acee-30c9022b9170","Type":"ContainerDied","Data":"e0bb131dcce7694c9a63e34597dce47a19e480e2578bc471243f2024f6cd4bca"} Nov 24 07:08:16 crc kubenswrapper[4809]: I1124 07:08:16.084644 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mn5w6" Nov 24 07:08:16 crc kubenswrapper[4809]: I1124 07:08:16.084709 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mn5w6" Nov 24 07:08:16 crc kubenswrapper[4809]: I1124 07:08:16.144473 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mn5w6" Nov 24 07:08:16 crc kubenswrapper[4809]: I1124 07:08:16.906782 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e191709-50db-49f9-b360-4d0e6021caf7" path="/var/lib/kubelet/pods/4e191709-50db-49f9-b360-4d0e6021caf7/volumes" Nov 24 07:08:17 crc kubenswrapper[4809]: I1124 07:08:17.028104 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mn5w6" Nov 24 07:08:17 crc kubenswrapper[4809]: I1124 07:08:17.360104 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b" Nov 24 07:08:17 crc kubenswrapper[4809]: I1124 07:08:17.524185 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lk4qb\" (UniqueName: \"kubernetes.io/projected/ed47ae85-6bd2-4e58-acee-30c9022b9170-kube-api-access-lk4qb\") pod \"ed47ae85-6bd2-4e58-acee-30c9022b9170\" (UID: \"ed47ae85-6bd2-4e58-acee-30c9022b9170\") " Nov 24 07:08:17 crc kubenswrapper[4809]: I1124 07:08:17.524354 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ed47ae85-6bd2-4e58-acee-30c9022b9170-util\") pod \"ed47ae85-6bd2-4e58-acee-30c9022b9170\" (UID: \"ed47ae85-6bd2-4e58-acee-30c9022b9170\") " Nov 24 07:08:17 crc kubenswrapper[4809]: I1124 07:08:17.524467 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ed47ae85-6bd2-4e58-acee-30c9022b9170-bundle\") pod \"ed47ae85-6bd2-4e58-acee-30c9022b9170\" (UID: \"ed47ae85-6bd2-4e58-acee-30c9022b9170\") " Nov 24 07:08:17 crc kubenswrapper[4809]: I1124 07:08:17.525642 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed47ae85-6bd2-4e58-acee-30c9022b9170-bundle" (OuterVolumeSpecName: "bundle") pod "ed47ae85-6bd2-4e58-acee-30c9022b9170" (UID: "ed47ae85-6bd2-4e58-acee-30c9022b9170"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:08:17 crc kubenswrapper[4809]: I1124 07:08:17.533924 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed47ae85-6bd2-4e58-acee-30c9022b9170-kube-api-access-lk4qb" (OuterVolumeSpecName: "kube-api-access-lk4qb") pod "ed47ae85-6bd2-4e58-acee-30c9022b9170" (UID: "ed47ae85-6bd2-4e58-acee-30c9022b9170"). InnerVolumeSpecName "kube-api-access-lk4qb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:08:17 crc kubenswrapper[4809]: I1124 07:08:17.540538 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed47ae85-6bd2-4e58-acee-30c9022b9170-util" (OuterVolumeSpecName: "util") pod "ed47ae85-6bd2-4e58-acee-30c9022b9170" (UID: "ed47ae85-6bd2-4e58-acee-30c9022b9170"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:08:17 crc kubenswrapper[4809]: I1124 07:08:17.626584 4809 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ed47ae85-6bd2-4e58-acee-30c9022b9170-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:17 crc kubenswrapper[4809]: I1124 07:08:17.626643 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lk4qb\" (UniqueName: \"kubernetes.io/projected/ed47ae85-6bd2-4e58-acee-30c9022b9170-kube-api-access-lk4qb\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:17 crc kubenswrapper[4809]: I1124 07:08:17.626666 4809 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ed47ae85-6bd2-4e58-acee-30c9022b9170-util\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:17 crc kubenswrapper[4809]: I1124 07:08:17.990594 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b" event={"ID":"ed47ae85-6bd2-4e58-acee-30c9022b9170","Type":"ContainerDied","Data":"ff60b9021d3741bcf16056a09e2a04eb6c956914ac7deec20fc249378c3357d5"} Nov 24 07:08:17 crc kubenswrapper[4809]: I1124 07:08:17.990685 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff60b9021d3741bcf16056a09e2a04eb6c956914ac7deec20fc249378c3357d5" Nov 24 07:08:17 crc kubenswrapper[4809]: I1124 07:08:17.990639 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b" Nov 24 07:08:21 crc kubenswrapper[4809]: I1124 07:08:21.316401 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mn5w6"] Nov 24 07:08:21 crc kubenswrapper[4809]: I1124 07:08:21.318325 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mn5w6" podUID="d60f4600-cb92-4b50-a958-caba8079a0fb" containerName="registry-server" containerID="cri-o://2afd482ef9ba62f193fa2c16f796f33ba0e88cb99591ee995fb5d690f343a440" gracePeriod=2 Nov 24 07:08:21 crc kubenswrapper[4809]: I1124 07:08:21.725037 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mn5w6" Nov 24 07:08:21 crc kubenswrapper[4809]: I1124 07:08:21.883433 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lbt9\" (UniqueName: \"kubernetes.io/projected/d60f4600-cb92-4b50-a958-caba8079a0fb-kube-api-access-4lbt9\") pod \"d60f4600-cb92-4b50-a958-caba8079a0fb\" (UID: \"d60f4600-cb92-4b50-a958-caba8079a0fb\") " Nov 24 07:08:21 crc kubenswrapper[4809]: I1124 07:08:21.883523 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d60f4600-cb92-4b50-a958-caba8079a0fb-utilities\") pod \"d60f4600-cb92-4b50-a958-caba8079a0fb\" (UID: \"d60f4600-cb92-4b50-a958-caba8079a0fb\") " Nov 24 07:08:21 crc kubenswrapper[4809]: I1124 07:08:21.883563 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d60f4600-cb92-4b50-a958-caba8079a0fb-catalog-content\") pod \"d60f4600-cb92-4b50-a958-caba8079a0fb\" (UID: \"d60f4600-cb92-4b50-a958-caba8079a0fb\") " Nov 24 07:08:21 crc kubenswrapper[4809]: I1124 07:08:21.884661 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d60f4600-cb92-4b50-a958-caba8079a0fb-utilities" (OuterVolumeSpecName: "utilities") pod "d60f4600-cb92-4b50-a958-caba8079a0fb" (UID: "d60f4600-cb92-4b50-a958-caba8079a0fb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:08:21 crc kubenswrapper[4809]: I1124 07:08:21.891882 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d60f4600-cb92-4b50-a958-caba8079a0fb-kube-api-access-4lbt9" (OuterVolumeSpecName: "kube-api-access-4lbt9") pod "d60f4600-cb92-4b50-a958-caba8079a0fb" (UID: "d60f4600-cb92-4b50-a958-caba8079a0fb"). InnerVolumeSpecName "kube-api-access-4lbt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:08:21 crc kubenswrapper[4809]: I1124 07:08:21.956243 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d60f4600-cb92-4b50-a958-caba8079a0fb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d60f4600-cb92-4b50-a958-caba8079a0fb" (UID: "d60f4600-cb92-4b50-a958-caba8079a0fb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:08:21 crc kubenswrapper[4809]: I1124 07:08:21.984676 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lbt9\" (UniqueName: \"kubernetes.io/projected/d60f4600-cb92-4b50-a958-caba8079a0fb-kube-api-access-4lbt9\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:21 crc kubenswrapper[4809]: I1124 07:08:21.984743 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d60f4600-cb92-4b50-a958-caba8079a0fb-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:21 crc kubenswrapper[4809]: I1124 07:08:21.984758 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d60f4600-cb92-4b50-a958-caba8079a0fb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:22 crc kubenswrapper[4809]: I1124 07:08:22.063947 4809 generic.go:334] "Generic (PLEG): container finished" podID="d60f4600-cb92-4b50-a958-caba8079a0fb" containerID="2afd482ef9ba62f193fa2c16f796f33ba0e88cb99591ee995fb5d690f343a440" exitCode=0 Nov 24 07:08:22 crc kubenswrapper[4809]: I1124 07:08:22.064019 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mn5w6" event={"ID":"d60f4600-cb92-4b50-a958-caba8079a0fb","Type":"ContainerDied","Data":"2afd482ef9ba62f193fa2c16f796f33ba0e88cb99591ee995fb5d690f343a440"} Nov 24 07:08:22 crc kubenswrapper[4809]: I1124 07:08:22.064045 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mn5w6" event={"ID":"d60f4600-cb92-4b50-a958-caba8079a0fb","Type":"ContainerDied","Data":"44b5a2c81bc458e88fd3dcc346fd81ae2cfa1fa888ad992fc963aced6f4e9f11"} Nov 24 07:08:22 crc kubenswrapper[4809]: I1124 07:08:22.064103 4809 scope.go:117] "RemoveContainer" containerID="2afd482ef9ba62f193fa2c16f796f33ba0e88cb99591ee995fb5d690f343a440" Nov 24 07:08:22 crc kubenswrapper[4809]: I1124 07:08:22.064128 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mn5w6" Nov 24 07:08:22 crc kubenswrapper[4809]: I1124 07:08:22.084533 4809 scope.go:117] "RemoveContainer" containerID="ab769e086a07620a47d65fb0394e990ce13ea4f06c5c584f47fbd8195ead7306" Nov 24 07:08:22 crc kubenswrapper[4809]: I1124 07:08:22.107311 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mn5w6"] Nov 24 07:08:22 crc kubenswrapper[4809]: I1124 07:08:22.113710 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mn5w6"] Nov 24 07:08:22 crc kubenswrapper[4809]: I1124 07:08:22.120548 4809 scope.go:117] "RemoveContainer" containerID="194ce5401fedba62813c939d956179a7ae74745293b273230a72f946eb2fd624" Nov 24 07:08:22 crc kubenswrapper[4809]: I1124 07:08:22.142262 4809 scope.go:117] "RemoveContainer" containerID="2afd482ef9ba62f193fa2c16f796f33ba0e88cb99591ee995fb5d690f343a440" Nov 24 07:08:22 crc kubenswrapper[4809]: E1124 07:08:22.142677 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2afd482ef9ba62f193fa2c16f796f33ba0e88cb99591ee995fb5d690f343a440\": container with ID starting with 2afd482ef9ba62f193fa2c16f796f33ba0e88cb99591ee995fb5d690f343a440 not found: ID does not exist" containerID="2afd482ef9ba62f193fa2c16f796f33ba0e88cb99591ee995fb5d690f343a440" Nov 24 07:08:22 crc kubenswrapper[4809]: I1124 07:08:22.142714 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2afd482ef9ba62f193fa2c16f796f33ba0e88cb99591ee995fb5d690f343a440"} err="failed to get container status \"2afd482ef9ba62f193fa2c16f796f33ba0e88cb99591ee995fb5d690f343a440\": rpc error: code = NotFound desc = could not find container \"2afd482ef9ba62f193fa2c16f796f33ba0e88cb99591ee995fb5d690f343a440\": container with ID starting with 2afd482ef9ba62f193fa2c16f796f33ba0e88cb99591ee995fb5d690f343a440 not found: ID does not exist" Nov 24 07:08:22 crc kubenswrapper[4809]: I1124 07:08:22.142738 4809 scope.go:117] "RemoveContainer" containerID="ab769e086a07620a47d65fb0394e990ce13ea4f06c5c584f47fbd8195ead7306" Nov 24 07:08:22 crc kubenswrapper[4809]: E1124 07:08:22.143044 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab769e086a07620a47d65fb0394e990ce13ea4f06c5c584f47fbd8195ead7306\": container with ID starting with ab769e086a07620a47d65fb0394e990ce13ea4f06c5c584f47fbd8195ead7306 not found: ID does not exist" containerID="ab769e086a07620a47d65fb0394e990ce13ea4f06c5c584f47fbd8195ead7306" Nov 24 07:08:22 crc kubenswrapper[4809]: I1124 07:08:22.143062 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab769e086a07620a47d65fb0394e990ce13ea4f06c5c584f47fbd8195ead7306"} err="failed to get container status \"ab769e086a07620a47d65fb0394e990ce13ea4f06c5c584f47fbd8195ead7306\": rpc error: code = NotFound desc = could not find container \"ab769e086a07620a47d65fb0394e990ce13ea4f06c5c584f47fbd8195ead7306\": container with ID starting with ab769e086a07620a47d65fb0394e990ce13ea4f06c5c584f47fbd8195ead7306 not found: ID does not exist" Nov 24 07:08:22 crc kubenswrapper[4809]: I1124 07:08:22.143075 4809 scope.go:117] "RemoveContainer" containerID="194ce5401fedba62813c939d956179a7ae74745293b273230a72f946eb2fd624" Nov 24 07:08:22 crc kubenswrapper[4809]: E1124 07:08:22.143296 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"194ce5401fedba62813c939d956179a7ae74745293b273230a72f946eb2fd624\": container with ID starting with 194ce5401fedba62813c939d956179a7ae74745293b273230a72f946eb2fd624 not found: ID does not exist" containerID="194ce5401fedba62813c939d956179a7ae74745293b273230a72f946eb2fd624" Nov 24 07:08:22 crc kubenswrapper[4809]: I1124 07:08:22.143312 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"194ce5401fedba62813c939d956179a7ae74745293b273230a72f946eb2fd624"} err="failed to get container status \"194ce5401fedba62813c939d956179a7ae74745293b273230a72f946eb2fd624\": rpc error: code = NotFound desc = could not find container \"194ce5401fedba62813c939d956179a7ae74745293b273230a72f946eb2fd624\": container with ID starting with 194ce5401fedba62813c939d956179a7ae74745293b273230a72f946eb2fd624 not found: ID does not exist" Nov 24 07:08:22 crc kubenswrapper[4809]: I1124 07:08:22.900392 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d60f4600-cb92-4b50-a958-caba8079a0fb" path="/var/lib/kubelet/pods/d60f4600-cb92-4b50-a958-caba8079a0fb/volumes" Nov 24 07:08:23 crc kubenswrapper[4809]: I1124 07:08:23.918706 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6c94b45f4d-pdjb4"] Nov 24 07:08:23 crc kubenswrapper[4809]: E1124 07:08:23.918949 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80dfda6c-ae33-4dbd-9750-67079eeacffc" containerName="registry-server" Nov 24 07:08:23 crc kubenswrapper[4809]: I1124 07:08:23.918982 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="80dfda6c-ae33-4dbd-9750-67079eeacffc" containerName="registry-server" Nov 24 07:08:23 crc kubenswrapper[4809]: E1124 07:08:23.918998 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d60f4600-cb92-4b50-a958-caba8079a0fb" containerName="registry-server" Nov 24 07:08:23 crc kubenswrapper[4809]: I1124 07:08:23.919007 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d60f4600-cb92-4b50-a958-caba8079a0fb" containerName="registry-server" Nov 24 07:08:23 crc kubenswrapper[4809]: E1124 07:08:23.919016 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed47ae85-6bd2-4e58-acee-30c9022b9170" containerName="util" Nov 24 07:08:23 crc kubenswrapper[4809]: I1124 07:08:23.919023 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed47ae85-6bd2-4e58-acee-30c9022b9170" containerName="util" Nov 24 07:08:23 crc kubenswrapper[4809]: E1124 07:08:23.919033 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e191709-50db-49f9-b360-4d0e6021caf7" containerName="extract-utilities" Nov 24 07:08:23 crc kubenswrapper[4809]: I1124 07:08:23.919042 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e191709-50db-49f9-b360-4d0e6021caf7" containerName="extract-utilities" Nov 24 07:08:23 crc kubenswrapper[4809]: E1124 07:08:23.919055 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d60f4600-cb92-4b50-a958-caba8079a0fb" containerName="extract-utilities" Nov 24 07:08:23 crc kubenswrapper[4809]: I1124 07:08:23.919062 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d60f4600-cb92-4b50-a958-caba8079a0fb" containerName="extract-utilities" Nov 24 07:08:23 crc kubenswrapper[4809]: E1124 07:08:23.919072 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e191709-50db-49f9-b360-4d0e6021caf7" containerName="extract-content" Nov 24 07:08:23 crc kubenswrapper[4809]: I1124 07:08:23.919080 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e191709-50db-49f9-b360-4d0e6021caf7" containerName="extract-content" Nov 24 07:08:23 crc kubenswrapper[4809]: E1124 07:08:23.919091 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed47ae85-6bd2-4e58-acee-30c9022b9170" containerName="pull" Nov 24 07:08:23 crc kubenswrapper[4809]: I1124 07:08:23.919098 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed47ae85-6bd2-4e58-acee-30c9022b9170" containerName="pull" Nov 24 07:08:23 crc kubenswrapper[4809]: E1124 07:08:23.919111 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80dfda6c-ae33-4dbd-9750-67079eeacffc" containerName="extract-utilities" Nov 24 07:08:23 crc kubenswrapper[4809]: I1124 07:08:23.919118 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="80dfda6c-ae33-4dbd-9750-67079eeacffc" containerName="extract-utilities" Nov 24 07:08:23 crc kubenswrapper[4809]: E1124 07:08:23.919129 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed47ae85-6bd2-4e58-acee-30c9022b9170" containerName="extract" Nov 24 07:08:23 crc kubenswrapper[4809]: I1124 07:08:23.919136 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed47ae85-6bd2-4e58-acee-30c9022b9170" containerName="extract" Nov 24 07:08:23 crc kubenswrapper[4809]: E1124 07:08:23.919149 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d60f4600-cb92-4b50-a958-caba8079a0fb" containerName="extract-content" Nov 24 07:08:23 crc kubenswrapper[4809]: I1124 07:08:23.919156 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d60f4600-cb92-4b50-a958-caba8079a0fb" containerName="extract-content" Nov 24 07:08:23 crc kubenswrapper[4809]: E1124 07:08:23.919168 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e191709-50db-49f9-b360-4d0e6021caf7" containerName="registry-server" Nov 24 07:08:23 crc kubenswrapper[4809]: I1124 07:08:23.919175 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e191709-50db-49f9-b360-4d0e6021caf7" containerName="registry-server" Nov 24 07:08:23 crc kubenswrapper[4809]: E1124 07:08:23.919186 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80dfda6c-ae33-4dbd-9750-67079eeacffc" containerName="extract-content" Nov 24 07:08:23 crc kubenswrapper[4809]: I1124 07:08:23.919193 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="80dfda6c-ae33-4dbd-9750-67079eeacffc" containerName="extract-content" Nov 24 07:08:23 crc kubenswrapper[4809]: I1124 07:08:23.919333 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d60f4600-cb92-4b50-a958-caba8079a0fb" containerName="registry-server" Nov 24 07:08:23 crc kubenswrapper[4809]: I1124 07:08:23.919350 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="80dfda6c-ae33-4dbd-9750-67079eeacffc" containerName="registry-server" Nov 24 07:08:23 crc kubenswrapper[4809]: I1124 07:08:23.919362 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e191709-50db-49f9-b360-4d0e6021caf7" containerName="registry-server" Nov 24 07:08:23 crc kubenswrapper[4809]: I1124 07:08:23.919376 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed47ae85-6bd2-4e58-acee-30c9022b9170" containerName="extract" Nov 24 07:08:23 crc kubenswrapper[4809]: I1124 07:08:23.920048 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6c94b45f4d-pdjb4" Nov 24 07:08:23 crc kubenswrapper[4809]: I1124 07:08:23.921733 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-b2kwt" Nov 24 07:08:23 crc kubenswrapper[4809]: I1124 07:08:23.948600 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6c94b45f4d-pdjb4"] Nov 24 07:08:24 crc kubenswrapper[4809]: I1124 07:08:24.017729 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4jpn\" (UniqueName: \"kubernetes.io/projected/a9cc582d-d6ce-4bb6-b698-9733ce8be925-kube-api-access-f4jpn\") pod \"openstack-operator-controller-operator-6c94b45f4d-pdjb4\" (UID: \"a9cc582d-d6ce-4bb6-b698-9733ce8be925\") " pod="openstack-operators/openstack-operator-controller-operator-6c94b45f4d-pdjb4" Nov 24 07:08:24 crc kubenswrapper[4809]: I1124 07:08:24.118883 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4jpn\" (UniqueName: \"kubernetes.io/projected/a9cc582d-d6ce-4bb6-b698-9733ce8be925-kube-api-access-f4jpn\") pod \"openstack-operator-controller-operator-6c94b45f4d-pdjb4\" (UID: \"a9cc582d-d6ce-4bb6-b698-9733ce8be925\") " pod="openstack-operators/openstack-operator-controller-operator-6c94b45f4d-pdjb4" Nov 24 07:08:24 crc kubenswrapper[4809]: I1124 07:08:24.137125 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4jpn\" (UniqueName: \"kubernetes.io/projected/a9cc582d-d6ce-4bb6-b698-9733ce8be925-kube-api-access-f4jpn\") pod \"openstack-operator-controller-operator-6c94b45f4d-pdjb4\" (UID: \"a9cc582d-d6ce-4bb6-b698-9733ce8be925\") " pod="openstack-operators/openstack-operator-controller-operator-6c94b45f4d-pdjb4" Nov 24 07:08:24 crc kubenswrapper[4809]: I1124 07:08:24.236044 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6c94b45f4d-pdjb4" Nov 24 07:08:24 crc kubenswrapper[4809]: I1124 07:08:24.693569 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6c94b45f4d-pdjb4"] Nov 24 07:08:25 crc kubenswrapper[4809]: I1124 07:08:25.085679 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6c94b45f4d-pdjb4" event={"ID":"a9cc582d-d6ce-4bb6-b698-9733ce8be925","Type":"ContainerStarted","Data":"41389dac890935d15f8a24e70eb98ffa3a3ee014e6f80b4dd308010dad56b0ba"} Nov 24 07:08:31 crc kubenswrapper[4809]: I1124 07:08:31.130192 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6c94b45f4d-pdjb4" event={"ID":"a9cc582d-d6ce-4bb6-b698-9733ce8be925","Type":"ContainerStarted","Data":"51b746dfa2fdb48e87787c38c083b66bc7133782748f94fb54809343748c57ad"} Nov 24 07:08:33 crc kubenswrapper[4809]: I1124 07:08:33.148762 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6c94b45f4d-pdjb4" event={"ID":"a9cc582d-d6ce-4bb6-b698-9733ce8be925","Type":"ContainerStarted","Data":"4e8eaf6befcc5ca39cacee28197c55677435f990f3bb0d8cf4d00a7fe014fe56"} Nov 24 07:08:33 crc kubenswrapper[4809]: I1124 07:08:33.149918 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6c94b45f4d-pdjb4" Nov 24 07:08:33 crc kubenswrapper[4809]: I1124 07:08:33.189681 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6c94b45f4d-pdjb4" podStartSLOduration=2.559338756 podStartE2EDuration="10.189662846s" podCreationTimestamp="2025-11-24 07:08:23 +0000 UTC" firstStartedPulling="2025-11-24 07:08:24.698008353 +0000 UTC m=+804.598600058" lastFinishedPulling="2025-11-24 07:08:32.328332403 +0000 UTC m=+812.228924148" observedRunningTime="2025-11-24 07:08:33.185579763 +0000 UTC m=+813.086171508" watchObservedRunningTime="2025-11-24 07:08:33.189662846 +0000 UTC m=+813.090254561" Nov 24 07:08:44 crc kubenswrapper[4809]: I1124 07:08:44.238157 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6c94b45f4d-pdjb4" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.524922 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-6k9xr"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.526232 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-6k9xr" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.527694 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-74rsq" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.535213 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-49dpw"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.536453 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-49dpw" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.537881 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-g56hl" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.554710 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-49dpw"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.562271 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-plhlb"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.563386 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-plhlb" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.564895 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-qr229" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.572443 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-plhlb"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.580873 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-6k9xr"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.594172 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-8nzmk"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.595044 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7969689c84-8nzmk" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.596639 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-wkb7c" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.611167 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-8nzmk"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.617358 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-4ddch"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.619008 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-4ddch" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.622589 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-mpjbl" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.632541 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-xf4fh"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.633498 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-xf4fh" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.635384 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-4ddch"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.637383 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-r97fk" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.642886 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-xf4fh"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.678568 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-4vrfb"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.679592 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-4vrfb" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.683261 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-vrbs5" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.683411 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.687654 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-jfjqt"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.689194 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jfjqt" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.691653 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-fnf6s" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.692478 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-4vrfb"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.709608 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-wh5fq"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.711523 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-wh5fq" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.714023 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-mtj4k" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.718031 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-jfjqt"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.721235 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn4hx\" (UniqueName: \"kubernetes.io/projected/e525562e-b6b2-415f-b6e7-1c0bfde8a56f-kube-api-access-sn4hx\") pod \"cinder-operator-controller-manager-6498cbf48f-49dpw\" (UID: \"e525562e-b6b2-415f-b6e7-1c0bfde8a56f\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-49dpw" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.721277 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9l6h\" (UniqueName: \"kubernetes.io/projected/0a1424d6-9ecf-4a06-b174-15d13f61b916-kube-api-access-c9l6h\") pod \"glance-operator-controller-manager-7969689c84-8nzmk\" (UID: \"0a1424d6-9ecf-4a06-b174-15d13f61b916\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-8nzmk" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.721316 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntw2w\" (UniqueName: \"kubernetes.io/projected/bfcf902b-83f0-48f9-ad6b-eb7963f43af4-kube-api-access-ntw2w\") pod \"barbican-operator-controller-manager-75fb479bcc-6k9xr\" (UID: \"bfcf902b-83f0-48f9-ad6b-eb7963f43af4\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-6k9xr" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.721360 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dg9x\" (UniqueName: \"kubernetes.io/projected/200f6896-3100-46ea-94f4-69637762a259-kube-api-access-8dg9x\") pod \"designate-operator-controller-manager-767ccfd65f-plhlb\" (UID: \"200f6896-3100-46ea-94f4-69637762a259\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-plhlb" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.721379 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqtw4\" (UniqueName: \"kubernetes.io/projected/d78cc4fe-bab9-4521-a1d2-ed6efd341789-kube-api-access-jqtw4\") pod \"heat-operator-controller-manager-56f54d6746-4ddch\" (UID: \"d78cc4fe-bab9-4521-a1d2-ed6efd341789\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-4ddch" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.738506 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-wh5fq"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.767273 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-rp7r4"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.777767 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-rp7r4" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.780796 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-qrq94" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.792183 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-bz8c8"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.822604 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58f887965d-bz8c8" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.830561 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpvd2\" (UniqueName: \"kubernetes.io/projected/b0a19ac9-d266-4416-8590-3c8de4dadc7f-kube-api-access-fpvd2\") pod \"keystone-operator-controller-manager-7454b96578-wh5fq\" (UID: \"b0a19ac9-d266-4416-8590-3c8de4dadc7f\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-wh5fq" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.830607 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw5bh\" (UniqueName: \"kubernetes.io/projected/5b90cc28-7c58-4826-b849-3456bb654f0d-kube-api-access-bw5bh\") pod \"infra-operator-controller-manager-6dd8864d7c-4vrfb\" (UID: \"5b90cc28-7c58-4826-b849-3456bb654f0d\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-4vrfb" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.831228 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9l6h\" (UniqueName: \"kubernetes.io/projected/0a1424d6-9ecf-4a06-b174-15d13f61b916-kube-api-access-c9l6h\") pod \"glance-operator-controller-manager-7969689c84-8nzmk\" (UID: \"0a1424d6-9ecf-4a06-b174-15d13f61b916\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-8nzmk" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.831309 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b90cc28-7c58-4826-b849-3456bb654f0d-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-4vrfb\" (UID: \"5b90cc28-7c58-4826-b849-3456bb654f0d\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-4vrfb" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.831331 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-686d7\" (UniqueName: \"kubernetes.io/projected/4d4aa35c-faa7-4b93-aecd-ee4645f11916-kube-api-access-686d7\") pod \"horizon-operator-controller-manager-598f69df5d-xf4fh\" (UID: \"4d4aa35c-faa7-4b93-aecd-ee4645f11916\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-xf4fh" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.831364 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntw2w\" (UniqueName: \"kubernetes.io/projected/bfcf902b-83f0-48f9-ad6b-eb7963f43af4-kube-api-access-ntw2w\") pod \"barbican-operator-controller-manager-75fb479bcc-6k9xr\" (UID: \"bfcf902b-83f0-48f9-ad6b-eb7963f43af4\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-6k9xr" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.831462 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fx4t7\" (UniqueName: \"kubernetes.io/projected/5f4f9ac0-6714-4f4a-a7fb-8d65b19c919c-kube-api-access-fx4t7\") pod \"ironic-operator-controller-manager-99b499f4-jfjqt\" (UID: \"5f4f9ac0-6714-4f4a-a7fb-8d65b19c919c\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jfjqt" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.831488 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dg9x\" (UniqueName: \"kubernetes.io/projected/200f6896-3100-46ea-94f4-69637762a259-kube-api-access-8dg9x\") pod \"designate-operator-controller-manager-767ccfd65f-plhlb\" (UID: \"200f6896-3100-46ea-94f4-69637762a259\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-plhlb" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.831510 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqtw4\" (UniqueName: \"kubernetes.io/projected/d78cc4fe-bab9-4521-a1d2-ed6efd341789-kube-api-access-jqtw4\") pod \"heat-operator-controller-manager-56f54d6746-4ddch\" (UID: \"d78cc4fe-bab9-4521-a1d2-ed6efd341789\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-4ddch" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.831549 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn4hx\" (UniqueName: \"kubernetes.io/projected/e525562e-b6b2-415f-b6e7-1c0bfde8a56f-kube-api-access-sn4hx\") pod \"cinder-operator-controller-manager-6498cbf48f-49dpw\" (UID: \"e525562e-b6b2-415f-b6e7-1c0bfde8a56f\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-49dpw" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.844689 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-wxfdf" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.855386 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn4hx\" (UniqueName: \"kubernetes.io/projected/e525562e-b6b2-415f-b6e7-1c0bfde8a56f-kube-api-access-sn4hx\") pod \"cinder-operator-controller-manager-6498cbf48f-49dpw\" (UID: \"e525562e-b6b2-415f-b6e7-1c0bfde8a56f\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-49dpw" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.860658 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-49dpw" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.863566 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9l6h\" (UniqueName: \"kubernetes.io/projected/0a1424d6-9ecf-4a06-b174-15d13f61b916-kube-api-access-c9l6h\") pod \"glance-operator-controller-manager-7969689c84-8nzmk\" (UID: \"0a1424d6-9ecf-4a06-b174-15d13f61b916\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-8nzmk" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.864153 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqtw4\" (UniqueName: \"kubernetes.io/projected/d78cc4fe-bab9-4521-a1d2-ed6efd341789-kube-api-access-jqtw4\") pod \"heat-operator-controller-manager-56f54d6746-4ddch\" (UID: \"d78cc4fe-bab9-4521-a1d2-ed6efd341789\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-4ddch" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.864344 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-rp7r4"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.882795 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-rgkd9"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.884572 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-rgkd9" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.884910 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dg9x\" (UniqueName: \"kubernetes.io/projected/200f6896-3100-46ea-94f4-69637762a259-kube-api-access-8dg9x\") pod \"designate-operator-controller-manager-767ccfd65f-plhlb\" (UID: \"200f6896-3100-46ea-94f4-69637762a259\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-plhlb" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.887546 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-twjx5" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.899590 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntw2w\" (UniqueName: \"kubernetes.io/projected/bfcf902b-83f0-48f9-ad6b-eb7963f43af4-kube-api-access-ntw2w\") pod \"barbican-operator-controller-manager-75fb479bcc-6k9xr\" (UID: \"bfcf902b-83f0-48f9-ad6b-eb7963f43af4\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-6k9xr" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.911796 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7969689c84-8nzmk" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.931313 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-bz8c8"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.932682 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fx4t7\" (UniqueName: \"kubernetes.io/projected/5f4f9ac0-6714-4f4a-a7fb-8d65b19c919c-kube-api-access-fx4t7\") pod \"ironic-operator-controller-manager-99b499f4-jfjqt\" (UID: \"5f4f9ac0-6714-4f4a-a7fb-8d65b19c919c\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jfjqt" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.934404 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-tk2hk"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.935055 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpvd2\" (UniqueName: \"kubernetes.io/projected/b0a19ac9-d266-4416-8590-3c8de4dadc7f-kube-api-access-fpvd2\") pod \"keystone-operator-controller-manager-7454b96578-wh5fq\" (UID: \"b0a19ac9-d266-4416-8590-3c8de4dadc7f\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-wh5fq" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.935111 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw5bh\" (UniqueName: \"kubernetes.io/projected/5b90cc28-7c58-4826-b849-3456bb654f0d-kube-api-access-bw5bh\") pod \"infra-operator-controller-manager-6dd8864d7c-4vrfb\" (UID: \"5b90cc28-7c58-4826-b849-3456bb654f0d\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-4vrfb" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.935209 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b90cc28-7c58-4826-b849-3456bb654f0d-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-4vrfb\" (UID: \"5b90cc28-7c58-4826-b849-3456bb654f0d\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-4vrfb" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.935237 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-686d7\" (UniqueName: \"kubernetes.io/projected/4d4aa35c-faa7-4b93-aecd-ee4645f11916-kube-api-access-686d7\") pod \"horizon-operator-controller-manager-598f69df5d-xf4fh\" (UID: \"4d4aa35c-faa7-4b93-aecd-ee4645f11916\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-xf4fh" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.935351 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gmmt\" (UniqueName: \"kubernetes.io/projected/2f364be5-58b1-4d57-97f5-65a13d63e34e-kube-api-access-7gmmt\") pod \"mariadb-operator-controller-manager-54b5986bb8-rp7r4\" (UID: \"2f364be5-58b1-4d57-97f5-65a13d63e34e\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-rp7r4" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.935385 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnnsz\" (UniqueName: \"kubernetes.io/projected/a2b969b3-07c7-4bf3-ab18-5317d8f0a372-kube-api-access-pnnsz\") pod \"manila-operator-controller-manager-58f887965d-bz8c8\" (UID: \"a2b969b3-07c7-4bf3-ab18-5317d8f0a372\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-bz8c8" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.936133 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-tk2hk" Nov 24 07:09:09 crc kubenswrapper[4809]: E1124 07:09:09.937190 4809 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 24 07:09:09 crc kubenswrapper[4809]: E1124 07:09:09.937259 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5b90cc28-7c58-4826-b849-3456bb654f0d-cert podName:5b90cc28-7c58-4826-b849-3456bb654f0d nodeName:}" failed. No retries permitted until 2025-11-24 07:09:10.437237676 +0000 UTC m=+850.337829381 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5b90cc28-7c58-4826-b849-3456bb654f0d-cert") pod "infra-operator-controller-manager-6dd8864d7c-4vrfb" (UID: "5b90cc28-7c58-4826-b849-3456bb654f0d") : secret "infra-operator-webhook-server-cert" not found Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.945601 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-zcgq2" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.946167 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-84nbm"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.947124 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-84nbm" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.951667 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-rgkd9"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.953123 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-jmlpf" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.962152 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-4ddch" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.962544 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-tk2hk"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.983984 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpvd2\" (UniqueName: \"kubernetes.io/projected/b0a19ac9-d266-4416-8590-3c8de4dadc7f-kube-api-access-fpvd2\") pod \"keystone-operator-controller-manager-7454b96578-wh5fq\" (UID: \"b0a19ac9-d266-4416-8590-3c8de4dadc7f\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-wh5fq" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.984680 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-686d7\" (UniqueName: \"kubernetes.io/projected/4d4aa35c-faa7-4b93-aecd-ee4645f11916-kube-api-access-686d7\") pod \"horizon-operator-controller-manager-598f69df5d-xf4fh\" (UID: \"4d4aa35c-faa7-4b93-aecd-ee4645f11916\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-xf4fh" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.988611 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-84nbm"] Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.989363 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fx4t7\" (UniqueName: \"kubernetes.io/projected/5f4f9ac0-6714-4f4a-a7fb-8d65b19c919c-kube-api-access-fx4t7\") pod \"ironic-operator-controller-manager-99b499f4-jfjqt\" (UID: \"5f4f9ac0-6714-4f4a-a7fb-8d65b19c919c\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jfjqt" Nov 24 07:09:09 crc kubenswrapper[4809]: I1124 07:09:09.993201 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw5bh\" (UniqueName: \"kubernetes.io/projected/5b90cc28-7c58-4826-b849-3456bb654f0d-kube-api-access-bw5bh\") pod \"infra-operator-controller-manager-6dd8864d7c-4vrfb\" (UID: \"5b90cc28-7c58-4826-b849-3456bb654f0d\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-4vrfb" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.007431 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-7vg75"] Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.008813 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-7vg75" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.014517 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-59brn"] Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.015778 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-59brn" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.016672 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.016913 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-p2rkz" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.021615 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-7vg75"] Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.024997 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-5jhxj" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.028490 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-rtxsf"] Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.036208 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rtxsf" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.041055 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-59brn"] Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.046477 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jfjqt" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.051208 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-5bx2f" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.052350 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-rtxsf"] Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.056557 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-mfgld"] Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.057046 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gmmt\" (UniqueName: \"kubernetes.io/projected/2f364be5-58b1-4d57-97f5-65a13d63e34e-kube-api-access-7gmmt\") pod \"mariadb-operator-controller-manager-54b5986bb8-rp7r4\" (UID: \"2f364be5-58b1-4d57-97f5-65a13d63e34e\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-rp7r4" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.057081 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnnsz\" (UniqueName: \"kubernetes.io/projected/a2b969b3-07c7-4bf3-ab18-5317d8f0a372-kube-api-access-pnnsz\") pod \"manila-operator-controller-manager-58f887965d-bz8c8\" (UID: \"a2b969b3-07c7-4bf3-ab18-5317d8f0a372\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-bz8c8" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.057127 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhcj8\" (UniqueName: \"kubernetes.io/projected/e94d989e-7aee-43d9-87c0-712355ca193f-kube-api-access-jhcj8\") pod \"neutron-operator-controller-manager-78bd47f458-rgkd9\" (UID: \"e94d989e-7aee-43d9-87c0-712355ca193f\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-rgkd9" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.057826 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d656998f4-mfgld" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.060319 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-mfgld"] Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.067287 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-bxwhc" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.073050 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4xjsj"] Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.075476 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnnsz\" (UniqueName: \"kubernetes.io/projected/a2b969b3-07c7-4bf3-ab18-5317d8f0a372-kube-api-access-pnnsz\") pod \"manila-operator-controller-manager-58f887965d-bz8c8\" (UID: \"a2b969b3-07c7-4bf3-ab18-5317d8f0a372\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-bz8c8" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.075516 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gmmt\" (UniqueName: \"kubernetes.io/projected/2f364be5-58b1-4d57-97f5-65a13d63e34e-kube-api-access-7gmmt\") pod \"mariadb-operator-controller-manager-54b5986bb8-rp7r4\" (UID: \"2f364be5-58b1-4d57-97f5-65a13d63e34e\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-rp7r4" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.079253 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4xjsj"] Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.079375 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4xjsj" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.079945 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-wh5fq" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.086741 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-4klxj" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.117025 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-wbzcn"] Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.134603 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-rp7r4" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.143054 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-wbzcn"] Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.143139 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-b4c496f69-wbzcn" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.144635 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-6k9xr" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.151357 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-22x9f" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.156229 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58f887965d-bz8c8" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.159047 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prlz5\" (UniqueName: \"kubernetes.io/projected/fd9ddd03-bc95-43e1-93ef-97ad5ca589b3-kube-api-access-prlz5\") pod \"nova-operator-controller-manager-cfbb9c588-tk2hk\" (UID: \"fd9ddd03-bc95-43e1-93ef-97ad5ca589b3\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-tk2hk" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.159096 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhcj8\" (UniqueName: \"kubernetes.io/projected/e94d989e-7aee-43d9-87c0-712355ca193f-kube-api-access-jhcj8\") pod \"neutron-operator-controller-manager-78bd47f458-rgkd9\" (UID: \"e94d989e-7aee-43d9-87c0-712355ca193f\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-rgkd9" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.159121 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzpnn\" (UniqueName: \"kubernetes.io/projected/2e599303-b44b-4ce5-9ee9-1e1c13ca856f-kube-api-access-pzpnn\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-7vg75\" (UID: \"2e599303-b44b-4ce5-9ee9-1e1c13ca856f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-7vg75" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.159174 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnzhs\" (UniqueName: \"kubernetes.io/projected/6269a5dc-fd2a-4fdd-9b18-4e505242efca-kube-api-access-gnzhs\") pod \"swift-operator-controller-manager-d656998f4-mfgld\" (UID: \"6269a5dc-fd2a-4fdd-9b18-4e505242efca\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-mfgld" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.159219 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2e599303-b44b-4ce5-9ee9-1e1c13ca856f-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-7vg75\" (UID: \"2e599303-b44b-4ce5-9ee9-1e1c13ca856f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-7vg75" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.159239 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkvbk\" (UniqueName: \"kubernetes.io/projected/9eeb9814-10f8-478d-888f-5cf058c99c4b-kube-api-access-tkvbk\") pod \"ovn-operator-controller-manager-54fc5f65b7-59brn\" (UID: \"9eeb9814-10f8-478d-888f-5cf058c99c4b\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-59brn" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.159258 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s58fs\" (UniqueName: \"kubernetes.io/projected/483b5843-7751-465a-9412-dbb15ef5145f-kube-api-access-s58fs\") pod \"placement-operator-controller-manager-5b797b8dff-rtxsf\" (UID: \"483b5843-7751-465a-9412-dbb15ef5145f\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rtxsf" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.159278 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr5w6\" (UniqueName: \"kubernetes.io/projected/bfe0f53e-685b-4cef-b763-90399c5fdaac-kube-api-access-wr5w6\") pod \"octavia-operator-controller-manager-54cfbf4c7d-84nbm\" (UID: \"bfe0f53e-685b-4cef-b763-90399c5fdaac\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-84nbm" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.177186 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhcj8\" (UniqueName: \"kubernetes.io/projected/e94d989e-7aee-43d9-87c0-712355ca193f-kube-api-access-jhcj8\") pod \"neutron-operator-controller-manager-78bd47f458-rgkd9\" (UID: \"e94d989e-7aee-43d9-87c0-712355ca193f\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-rgkd9" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.177647 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-plhlb" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.245203 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-twjz4"] Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.247719 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-twjz4" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.252790 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-jwj22" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.256221 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-xf4fh" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.264063 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnzhs\" (UniqueName: \"kubernetes.io/projected/6269a5dc-fd2a-4fdd-9b18-4e505242efca-kube-api-access-gnzhs\") pod \"swift-operator-controller-manager-d656998f4-mfgld\" (UID: \"6269a5dc-fd2a-4fdd-9b18-4e505242efca\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-mfgld" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.272167 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2e599303-b44b-4ce5-9ee9-1e1c13ca856f-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-7vg75\" (UID: \"2e599303-b44b-4ce5-9ee9-1e1c13ca856f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-7vg75" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.272216 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prknl\" (UniqueName: \"kubernetes.io/projected/39fc255b-7ac9-40b4-98e0-74d1f76c2696-kube-api-access-prknl\") pod \"telemetry-operator-controller-manager-6d4bf84b58-4xjsj\" (UID: \"39fc255b-7ac9-40b4-98e0-74d1f76c2696\") " pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4xjsj" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.272239 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkvbk\" (UniqueName: \"kubernetes.io/projected/9eeb9814-10f8-478d-888f-5cf058c99c4b-kube-api-access-tkvbk\") pod \"ovn-operator-controller-manager-54fc5f65b7-59brn\" (UID: \"9eeb9814-10f8-478d-888f-5cf058c99c4b\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-59brn" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.272270 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s58fs\" (UniqueName: \"kubernetes.io/projected/483b5843-7751-465a-9412-dbb15ef5145f-kube-api-access-s58fs\") pod \"placement-operator-controller-manager-5b797b8dff-rtxsf\" (UID: \"483b5843-7751-465a-9412-dbb15ef5145f\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rtxsf" Nov 24 07:09:10 crc kubenswrapper[4809]: E1124 07:09:10.273451 4809 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 07:09:10 crc kubenswrapper[4809]: E1124 07:09:10.273517 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e599303-b44b-4ce5-9ee9-1e1c13ca856f-cert podName:2e599303-b44b-4ce5-9ee9-1e1c13ca856f nodeName:}" failed. No retries permitted until 2025-11-24 07:09:10.773492375 +0000 UTC m=+850.674084080 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2e599303-b44b-4ce5-9ee9-1e1c13ca856f-cert") pod "openstack-baremetal-operator-controller-manager-8c7444f48-7vg75" (UID: "2e599303-b44b-4ce5-9ee9-1e1c13ca856f") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.282854 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-rgkd9" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.272299 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr5w6\" (UniqueName: \"kubernetes.io/projected/bfe0f53e-685b-4cef-b763-90399c5fdaac-kube-api-access-wr5w6\") pod \"octavia-operator-controller-manager-54cfbf4c7d-84nbm\" (UID: \"bfe0f53e-685b-4cef-b763-90399c5fdaac\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-84nbm" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.289461 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prlz5\" (UniqueName: \"kubernetes.io/projected/fd9ddd03-bc95-43e1-93ef-97ad5ca589b3-kube-api-access-prlz5\") pod \"nova-operator-controller-manager-cfbb9c588-tk2hk\" (UID: \"fd9ddd03-bc95-43e1-93ef-97ad5ca589b3\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-tk2hk" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.289541 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzpnn\" (UniqueName: \"kubernetes.io/projected/2e599303-b44b-4ce5-9ee9-1e1c13ca856f-kube-api-access-pzpnn\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-7vg75\" (UID: \"2e599303-b44b-4ce5-9ee9-1e1c13ca856f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-7vg75" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.289605 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfxpm\" (UniqueName: \"kubernetes.io/projected/cb240075-d846-4751-b026-6fa452a565a5-kube-api-access-kfxpm\") pod \"test-operator-controller-manager-b4c496f69-wbzcn\" (UID: \"cb240075-d846-4751-b026-6fa452a565a5\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-wbzcn" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.314774 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkvbk\" (UniqueName: \"kubernetes.io/projected/9eeb9814-10f8-478d-888f-5cf058c99c4b-kube-api-access-tkvbk\") pod \"ovn-operator-controller-manager-54fc5f65b7-59brn\" (UID: \"9eeb9814-10f8-478d-888f-5cf058c99c4b\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-59brn" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.323756 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s58fs\" (UniqueName: \"kubernetes.io/projected/483b5843-7751-465a-9412-dbb15ef5145f-kube-api-access-s58fs\") pod \"placement-operator-controller-manager-5b797b8dff-rtxsf\" (UID: \"483b5843-7751-465a-9412-dbb15ef5145f\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rtxsf" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.324738 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnzhs\" (UniqueName: \"kubernetes.io/projected/6269a5dc-fd2a-4fdd-9b18-4e505242efca-kube-api-access-gnzhs\") pod \"swift-operator-controller-manager-d656998f4-mfgld\" (UID: \"6269a5dc-fd2a-4fdd-9b18-4e505242efca\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-mfgld" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.327221 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-twjz4"] Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.333624 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzpnn\" (UniqueName: \"kubernetes.io/projected/2e599303-b44b-4ce5-9ee9-1e1c13ca856f-kube-api-access-pzpnn\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-7vg75\" (UID: \"2e599303-b44b-4ce5-9ee9-1e1c13ca856f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-7vg75" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.336013 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr5w6\" (UniqueName: \"kubernetes.io/projected/bfe0f53e-685b-4cef-b763-90399c5fdaac-kube-api-access-wr5w6\") pod \"octavia-operator-controller-manager-54cfbf4c7d-84nbm\" (UID: \"bfe0f53e-685b-4cef-b763-90399c5fdaac\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-84nbm" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.338887 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prlz5\" (UniqueName: \"kubernetes.io/projected/fd9ddd03-bc95-43e1-93ef-97ad5ca589b3-kube-api-access-prlz5\") pod \"nova-operator-controller-manager-cfbb9c588-tk2hk\" (UID: \"fd9ddd03-bc95-43e1-93ef-97ad5ca589b3\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-tk2hk" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.363262 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8fcb84d5c-xqd67"] Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.364465 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-8fcb84d5c-xqd67" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.370787 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-rtq6b" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.371355 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.373841 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-84nbm" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.392035 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8fcb84d5c-xqd67"] Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.394102 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hhkz\" (UniqueName: \"kubernetes.io/projected/df289b0d-b465-49d2-b1ee-4f0faf87b079-kube-api-access-2hhkz\") pod \"watcher-operator-controller-manager-8c6448b9f-twjz4\" (UID: \"df289b0d-b465-49d2-b1ee-4f0faf87b079\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-twjz4" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.394147 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prknl\" (UniqueName: \"kubernetes.io/projected/39fc255b-7ac9-40b4-98e0-74d1f76c2696-kube-api-access-prknl\") pod \"telemetry-operator-controller-manager-6d4bf84b58-4xjsj\" (UID: \"39fc255b-7ac9-40b4-98e0-74d1f76c2696\") " pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4xjsj" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.394218 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj22c\" (UniqueName: \"kubernetes.io/projected/5b38e6fb-1aef-4ecd-80ca-4665a9d3430a-kube-api-access-gj22c\") pod \"openstack-operator-controller-manager-8fcb84d5c-xqd67\" (UID: \"5b38e6fb-1aef-4ecd-80ca-4665a9d3430a\") " pod="openstack-operators/openstack-operator-controller-manager-8fcb84d5c-xqd67" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.394249 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfxpm\" (UniqueName: \"kubernetes.io/projected/cb240075-d846-4751-b026-6fa452a565a5-kube-api-access-kfxpm\") pod \"test-operator-controller-manager-b4c496f69-wbzcn\" (UID: \"cb240075-d846-4751-b026-6fa452a565a5\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-wbzcn" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.394266 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b38e6fb-1aef-4ecd-80ca-4665a9d3430a-cert\") pod \"openstack-operator-controller-manager-8fcb84d5c-xqd67\" (UID: \"5b38e6fb-1aef-4ecd-80ca-4665a9d3430a\") " pod="openstack-operators/openstack-operator-controller-manager-8fcb84d5c-xqd67" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.411926 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-px9vl"] Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.413242 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-px9vl" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.416332 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-h7mdw" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.422381 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-59brn" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.435349 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-px9vl"] Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.438650 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfxpm\" (UniqueName: \"kubernetes.io/projected/cb240075-d846-4751-b026-6fa452a565a5-kube-api-access-kfxpm\") pod \"test-operator-controller-manager-b4c496f69-wbzcn\" (UID: \"cb240075-d846-4751-b026-6fa452a565a5\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-wbzcn" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.445925 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prknl\" (UniqueName: \"kubernetes.io/projected/39fc255b-7ac9-40b4-98e0-74d1f76c2696-kube-api-access-prknl\") pod \"telemetry-operator-controller-manager-6d4bf84b58-4xjsj\" (UID: \"39fc255b-7ac9-40b4-98e0-74d1f76c2696\") " pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4xjsj" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.458507 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rtxsf" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.476299 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d656998f4-mfgld" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.479138 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4xjsj" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.490584 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-49dpw"] Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.502360 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj22c\" (UniqueName: \"kubernetes.io/projected/5b38e6fb-1aef-4ecd-80ca-4665a9d3430a-kube-api-access-gj22c\") pod \"openstack-operator-controller-manager-8fcb84d5c-xqd67\" (UID: \"5b38e6fb-1aef-4ecd-80ca-4665a9d3430a\") " pod="openstack-operators/openstack-operator-controller-manager-8fcb84d5c-xqd67" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.502486 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b38e6fb-1aef-4ecd-80ca-4665a9d3430a-cert\") pod \"openstack-operator-controller-manager-8fcb84d5c-xqd67\" (UID: \"5b38e6fb-1aef-4ecd-80ca-4665a9d3430a\") " pod="openstack-operators/openstack-operator-controller-manager-8fcb84d5c-xqd67" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.502564 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b90cc28-7c58-4826-b849-3456bb654f0d-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-4vrfb\" (UID: \"5b90cc28-7c58-4826-b849-3456bb654f0d\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-4vrfb" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.502600 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hhkz\" (UniqueName: \"kubernetes.io/projected/df289b0d-b465-49d2-b1ee-4f0faf87b079-kube-api-access-2hhkz\") pod \"watcher-operator-controller-manager-8c6448b9f-twjz4\" (UID: \"df289b0d-b465-49d2-b1ee-4f0faf87b079\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-twjz4" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.502643 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzhnm\" (UniqueName: \"kubernetes.io/projected/02a1aa7c-c3f7-4dcb-89ad-8d34f46ff4a1-kube-api-access-wzhnm\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-px9vl\" (UID: \"02a1aa7c-c3f7-4dcb-89ad-8d34f46ff4a1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-px9vl" Nov 24 07:09:10 crc kubenswrapper[4809]: E1124 07:09:10.503089 4809 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 24 07:09:10 crc kubenswrapper[4809]: E1124 07:09:10.503124 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5b38e6fb-1aef-4ecd-80ca-4665a9d3430a-cert podName:5b38e6fb-1aef-4ecd-80ca-4665a9d3430a nodeName:}" failed. No retries permitted until 2025-11-24 07:09:11.003112199 +0000 UTC m=+850.903703904 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5b38e6fb-1aef-4ecd-80ca-4665a9d3430a-cert") pod "openstack-operator-controller-manager-8fcb84d5c-xqd67" (UID: "5b38e6fb-1aef-4ecd-80ca-4665a9d3430a") : secret "webhook-server-cert" not found Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.506556 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b90cc28-7c58-4826-b849-3456bb654f0d-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-4vrfb\" (UID: \"5b90cc28-7c58-4826-b849-3456bb654f0d\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-4vrfb" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.507251 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-b4c496f69-wbzcn" Nov 24 07:09:10 crc kubenswrapper[4809]: W1124 07:09:10.529168 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode525562e_b6b2_415f_b6e7_1c0bfde8a56f.slice/crio-ac4a53d92db50aab7983e9462bd3d419894ad1ce46768b9a15c2ebc21295565e WatchSource:0}: Error finding container ac4a53d92db50aab7983e9462bd3d419894ad1ce46768b9a15c2ebc21295565e: Status 404 returned error can't find the container with id ac4a53d92db50aab7983e9462bd3d419894ad1ce46768b9a15c2ebc21295565e Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.561659 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj22c\" (UniqueName: \"kubernetes.io/projected/5b38e6fb-1aef-4ecd-80ca-4665a9d3430a-kube-api-access-gj22c\") pod \"openstack-operator-controller-manager-8fcb84d5c-xqd67\" (UID: \"5b38e6fb-1aef-4ecd-80ca-4665a9d3430a\") " pod="openstack-operators/openstack-operator-controller-manager-8fcb84d5c-xqd67" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.569817 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hhkz\" (UniqueName: \"kubernetes.io/projected/df289b0d-b465-49d2-b1ee-4f0faf87b079-kube-api-access-2hhkz\") pod \"watcher-operator-controller-manager-8c6448b9f-twjz4\" (UID: \"df289b0d-b465-49d2-b1ee-4f0faf87b079\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-twjz4" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.605665 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-twjz4" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.605936 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzhnm\" (UniqueName: \"kubernetes.io/projected/02a1aa7c-c3f7-4dcb-89ad-8d34f46ff4a1-kube-api-access-wzhnm\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-px9vl\" (UID: \"02a1aa7c-c3f7-4dcb-89ad-8d34f46ff4a1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-px9vl" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.620453 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-8nzmk"] Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.620898 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-4vrfb" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.625641 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-tk2hk" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.637825 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzhnm\" (UniqueName: \"kubernetes.io/projected/02a1aa7c-c3f7-4dcb-89ad-8d34f46ff4a1-kube-api-access-wzhnm\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-px9vl\" (UID: \"02a1aa7c-c3f7-4dcb-89ad-8d34f46ff4a1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-px9vl" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.685890 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-4ddch"] Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.695878 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-8nzmk" event={"ID":"0a1424d6-9ecf-4a06-b174-15d13f61b916","Type":"ContainerStarted","Data":"8c93411ec94d9c6b7b0e754d356dafb79de78ea85d7e3fe0a3eb5179fbe358f6"} Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.700999 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-49dpw" event={"ID":"e525562e-b6b2-415f-b6e7-1c0bfde8a56f","Type":"ContainerStarted","Data":"ac4a53d92db50aab7983e9462bd3d419894ad1ce46768b9a15c2ebc21295565e"} Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.779635 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-px9vl" Nov 24 07:09:10 crc kubenswrapper[4809]: W1124 07:09:10.784117 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd78cc4fe_bab9_4521_a1d2_ed6efd341789.slice/crio-0a9a63de7746aa82b108a01a9e0a4e4d316017053d1c4c98ca572f00fc2933c2 WatchSource:0}: Error finding container 0a9a63de7746aa82b108a01a9e0a4e4d316017053d1c4c98ca572f00fc2933c2: Status 404 returned error can't find the container with id 0a9a63de7746aa82b108a01a9e0a4e4d316017053d1c4c98ca572f00fc2933c2 Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.808339 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2e599303-b44b-4ce5-9ee9-1e1c13ca856f-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-7vg75\" (UID: \"2e599303-b44b-4ce5-9ee9-1e1c13ca856f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-7vg75" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.812634 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2e599303-b44b-4ce5-9ee9-1e1c13ca856f-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-7vg75\" (UID: \"2e599303-b44b-4ce5-9ee9-1e1c13ca856f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-7vg75" Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.860280 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-jfjqt"] Nov 24 07:09:10 crc kubenswrapper[4809]: I1124 07:09:10.861097 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-wh5fq"] Nov 24 07:09:10 crc kubenswrapper[4809]: W1124 07:09:10.972013 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f4f9ac0_6714_4f4a_a7fb_8d65b19c919c.slice/crio-8596ffd70e5f96a58e9506fb80b1f5f9df9997c3d2230c6487876c1503ca90c8 WatchSource:0}: Error finding container 8596ffd70e5f96a58e9506fb80b1f5f9df9997c3d2230c6487876c1503ca90c8: Status 404 returned error can't find the container with id 8596ffd70e5f96a58e9506fb80b1f5f9df9997c3d2230c6487876c1503ca90c8 Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.003277 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-7vg75" Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.020778 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b38e6fb-1aef-4ecd-80ca-4665a9d3430a-cert\") pod \"openstack-operator-controller-manager-8fcb84d5c-xqd67\" (UID: \"5b38e6fb-1aef-4ecd-80ca-4665a9d3430a\") " pod="openstack-operators/openstack-operator-controller-manager-8fcb84d5c-xqd67" Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.034193 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b38e6fb-1aef-4ecd-80ca-4665a9d3430a-cert\") pod \"openstack-operator-controller-manager-8fcb84d5c-xqd67\" (UID: \"5b38e6fb-1aef-4ecd-80ca-4665a9d3430a\") " pod="openstack-operators/openstack-operator-controller-manager-8fcb84d5c-xqd67" Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.044900 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-rp7r4"] Nov 24 07:09:11 crc kubenswrapper[4809]: W1124 07:09:11.067458 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f364be5_58b1_4d57_97f5_65a13d63e34e.slice/crio-d7bf05f33ce57041bab57ae6c70f3b6b004eb7651b504f6ef2d14202dfa234f3 WatchSource:0}: Error finding container d7bf05f33ce57041bab57ae6c70f3b6b004eb7651b504f6ef2d14202dfa234f3: Status 404 returned error can't find the container with id d7bf05f33ce57041bab57ae6c70f3b6b004eb7651b504f6ef2d14202dfa234f3 Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.133619 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-6k9xr"] Nov 24 07:09:11 crc kubenswrapper[4809]: W1124 07:09:11.156428 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbfcf902b_83f0_48f9_ad6b_eb7963f43af4.slice/crio-aa678da16ea528044b8e74eb0cdf0f77f8cc2a6403010cc05a364ba539370000 WatchSource:0}: Error finding container aa678da16ea528044b8e74eb0cdf0f77f8cc2a6403010cc05a364ba539370000: Status 404 returned error can't find the container with id aa678da16ea528044b8e74eb0cdf0f77f8cc2a6403010cc05a364ba539370000 Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.314612 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-8fcb84d5c-xqd67" Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.504538 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-plhlb"] Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.514898 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-84nbm"] Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.528231 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-bz8c8"] Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.534172 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-rgkd9"] Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.599094 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-xf4fh"] Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.603153 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-rtxsf"] Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.612489 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-mfgld"] Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.617814 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-twjz4"] Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.726411 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-bz8c8" event={"ID":"a2b969b3-07c7-4bf3-ab18-5317d8f0a372","Type":"ContainerStarted","Data":"ed481064aab5697e59ada3cdeedaa083db31ad9a0e84ea9c4c092fb14f361e41"} Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.727960 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-twjz4" event={"ID":"df289b0d-b465-49d2-b1ee-4f0faf87b079","Type":"ContainerStarted","Data":"e525659638b7d27a4d5f6ddf83c1de0568dc5e621c291767d0d2c3d5572dbe17"} Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.742070 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-wh5fq" event={"ID":"b0a19ac9-d266-4416-8590-3c8de4dadc7f","Type":"ContainerStarted","Data":"e4edd24ad70e2e76f6daa11529468056b9a228625df5aa640bb8f92ddba2d4e6"} Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.743400 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-rgkd9" event={"ID":"e94d989e-7aee-43d9-87c0-712355ca193f","Type":"ContainerStarted","Data":"bf9591676f4955da4edf9bc7b43069ce1ae89960a16c5825d7c39b1e5cbb5295"} Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.745625 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-6k9xr" event={"ID":"bfcf902b-83f0-48f9-ad6b-eb7963f43af4","Type":"ContainerStarted","Data":"aa678da16ea528044b8e74eb0cdf0f77f8cc2a6403010cc05a364ba539370000"} Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.747063 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-plhlb" event={"ID":"200f6896-3100-46ea-94f4-69637762a259","Type":"ContainerStarted","Data":"3592f76aa21b08dbc643055c7a1e3fc09d316105232dbdb54e37015913915061"} Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.748327 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jfjqt" event={"ID":"5f4f9ac0-6714-4f4a-a7fb-8d65b19c919c","Type":"ContainerStarted","Data":"8596ffd70e5f96a58e9506fb80b1f5f9df9997c3d2230c6487876c1503ca90c8"} Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.749474 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-4ddch" event={"ID":"d78cc4fe-bab9-4521-a1d2-ed6efd341789","Type":"ContainerStarted","Data":"0a9a63de7746aa82b108a01a9e0a4e4d316017053d1c4c98ca572f00fc2933c2"} Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.752358 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-84nbm" event={"ID":"bfe0f53e-685b-4cef-b763-90399c5fdaac","Type":"ContainerStarted","Data":"a79254907d6ea71817fa9fffce750d1ac74f9548c86cfced1d5724b762f9cdd3"} Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.754384 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-rp7r4" event={"ID":"2f364be5-58b1-4d57-97f5-65a13d63e34e","Type":"ContainerStarted","Data":"d7bf05f33ce57041bab57ae6c70f3b6b004eb7651b504f6ef2d14202dfa234f3"} Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.755537 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rtxsf" event={"ID":"483b5843-7751-465a-9412-dbb15ef5145f","Type":"ContainerStarted","Data":"a9e807552af543a5f68c3420b0f28acf783102c2461b8370fb069dfc7f58a89c"} Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.778406 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-xf4fh" event={"ID":"4d4aa35c-faa7-4b93-aecd-ee4645f11916","Type":"ContainerStarted","Data":"8124fb660b85fc437bba8a8c975b3459ed84b5c2e399f5ab90f5effcf19ae77f"} Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.780302 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-mfgld" event={"ID":"6269a5dc-fd2a-4fdd-9b18-4e505242efca","Type":"ContainerStarted","Data":"feffd7575f3d5794dae4d9bf27835f751a1362917c3b3579fdb4a4a1ae20dbc6"} Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.787730 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-tk2hk"] Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.794019 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-px9vl"] Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.800951 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-59brn"] Nov 24 07:09:11 crc kubenswrapper[4809]: W1124 07:09:11.801990 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd9ddd03_bc95_43e1_93ef_97ad5ca589b3.slice/crio-7ced5600ff67550fe7def3e72159eea4508e4c692d1977a87058be5de34a72c9 WatchSource:0}: Error finding container 7ced5600ff67550fe7def3e72159eea4508e4c692d1977a87058be5de34a72c9: Status 404 returned error can't find the container with id 7ced5600ff67550fe7def3e72159eea4508e4c692d1977a87058be5de34a72c9 Nov 24 07:09:11 crc kubenswrapper[4809]: W1124 07:09:11.811204 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02a1aa7c_c3f7_4dcb_89ad_8d34f46ff4a1.slice/crio-bf81f2f4045f227f284e56c26c9b0dcfc00f577628326da061bed19e37d4af12 WatchSource:0}: Error finding container bf81f2f4045f227f284e56c26c9b0dcfc00f577628326da061bed19e37d4af12: Status 404 returned error can't find the container with id bf81f2f4045f227f284e56c26c9b0dcfc00f577628326da061bed19e37d4af12 Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.815383 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-4vrfb"] Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.820284 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-wbzcn"] Nov 24 07:09:11 crc kubenswrapper[4809]: E1124 07:09:11.820497 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bw5bh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-6dd8864d7c-4vrfb_openstack-operators(5b90cc28-7c58-4826-b849-3456bb654f0d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.825122 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-7vg75"] Nov 24 07:09:11 crc kubenswrapper[4809]: E1124 07:09:11.828564 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tkvbk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-54fc5f65b7-59brn_openstack-operators(9eeb9814-10f8-478d-888f-5cf058c99c4b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 07:09:11 crc kubenswrapper[4809]: E1124 07:09:11.828604 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-prlz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-cfbb9c588-tk2hk_openstack-operators(fd9ddd03-bc95-43e1-93ef-97ad5ca589b3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.830895 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4xjsj"] Nov 24 07:09:11 crc kubenswrapper[4809]: W1124 07:09:11.832773 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb240075_d846_4751_b026_6fa452a565a5.slice/crio-a07ec5d31b5882da3a34737e5074a9c8e28cd855d36b12fc0ffa08db74c8a1f3 WatchSource:0}: Error finding container a07ec5d31b5882da3a34737e5074a9c8e28cd855d36b12fc0ffa08db74c8a1f3: Status 404 returned error can't find the container with id a07ec5d31b5882da3a34737e5074a9c8e28cd855d36b12fc0ffa08db74c8a1f3 Nov 24 07:09:11 crc kubenswrapper[4809]: E1124 07:09:11.837831 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kfxpm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-b4c496f69-wbzcn_openstack-operators(cb240075-d846-4751-b026-6fa452a565a5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 07:09:11 crc kubenswrapper[4809]: W1124 07:09:11.845316 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod39fc255b_7ac9_40b4_98e0_74d1f76c2696.slice/crio-3e9718f8d611bffe7d7fa17be91e3e79faec0e86d888c1a0776f001a4c1477a5 WatchSource:0}: Error finding container 3e9718f8d611bffe7d7fa17be91e3e79faec0e86d888c1a0776f001a4c1477a5: Status 404 returned error can't find the container with id 3e9718f8d611bffe7d7fa17be91e3e79faec0e86d888c1a0776f001a4c1477a5 Nov 24 07:09:11 crc kubenswrapper[4809]: I1124 07:09:11.846851 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8fcb84d5c-xqd67"] Nov 24 07:09:11 crc kubenswrapper[4809]: E1124 07:09:11.903469 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-prknl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-6d4bf84b58-4xjsj_openstack-operators(39fc255b-7ac9-40b4-98e0-74d1f76c2696): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 07:09:11 crc kubenswrapper[4809]: E1124 07:09:11.903669 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pzpnn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-8c7444f48-7vg75_openstack-operators(2e599303-b44b-4ce5-9ee9-1e1c13ca856f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 07:09:12 crc kubenswrapper[4809]: E1124 07:09:12.068885 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-4vrfb" podUID="5b90cc28-7c58-4826-b849-3456bb654f0d" Nov 24 07:09:12 crc kubenswrapper[4809]: E1124 07:09:12.202519 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-59brn" podUID="9eeb9814-10f8-478d-888f-5cf058c99c4b" Nov 24 07:09:12 crc kubenswrapper[4809]: E1124 07:09:12.286979 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-wbzcn" podUID="cb240075-d846-4751-b026-6fa452a565a5" Nov 24 07:09:12 crc kubenswrapper[4809]: E1124 07:09:12.304073 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-tk2hk" podUID="fd9ddd03-bc95-43e1-93ef-97ad5ca589b3" Nov 24 07:09:12 crc kubenswrapper[4809]: E1124 07:09:12.319116 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4xjsj" podUID="39fc255b-7ac9-40b4-98e0-74d1f76c2696" Nov 24 07:09:12 crc kubenswrapper[4809]: E1124 07:09:12.342646 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-7vg75" podUID="2e599303-b44b-4ce5-9ee9-1e1c13ca856f" Nov 24 07:09:12 crc kubenswrapper[4809]: I1124 07:09:12.799911 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-7vg75" event={"ID":"2e599303-b44b-4ce5-9ee9-1e1c13ca856f","Type":"ContainerStarted","Data":"9a970afd2d4999c4065115ef5f67a42e7c810f29b5d85f8805168f4b09fcefc9"} Nov 24 07:09:12 crc kubenswrapper[4809]: I1124 07:09:12.799956 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-7vg75" event={"ID":"2e599303-b44b-4ce5-9ee9-1e1c13ca856f","Type":"ContainerStarted","Data":"dd96a213b3a95148fa0ff621fe0b52a5c58d50940af2aaec572fee4dec831437"} Nov 24 07:09:12 crc kubenswrapper[4809]: E1124 07:09:12.802384 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-7vg75" podUID="2e599303-b44b-4ce5-9ee9-1e1c13ca856f" Nov 24 07:09:12 crc kubenswrapper[4809]: I1124 07:09:12.808572 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-59brn" event={"ID":"9eeb9814-10f8-478d-888f-5cf058c99c4b","Type":"ContainerStarted","Data":"fda020473a787ee5df67461244bf1c254f517c630edeeb39da39a1b5e606695f"} Nov 24 07:09:12 crc kubenswrapper[4809]: I1124 07:09:12.808616 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-59brn" event={"ID":"9eeb9814-10f8-478d-888f-5cf058c99c4b","Type":"ContainerStarted","Data":"2a55700cd444f76eef682588a300615b25a475c1f3745020f43cb40d6cb2031c"} Nov 24 07:09:12 crc kubenswrapper[4809]: E1124 07:09:12.810013 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-59brn" podUID="9eeb9814-10f8-478d-888f-5cf058c99c4b" Nov 24 07:09:12 crc kubenswrapper[4809]: I1124 07:09:12.839671 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4xjsj" event={"ID":"39fc255b-7ac9-40b4-98e0-74d1f76c2696","Type":"ContainerStarted","Data":"9a42cd7c4105eba5c9a4acda7d60f307ef8a532685d360a25d0b35d72e53e29f"} Nov 24 07:09:12 crc kubenswrapper[4809]: I1124 07:09:12.839708 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4xjsj" event={"ID":"39fc255b-7ac9-40b4-98e0-74d1f76c2696","Type":"ContainerStarted","Data":"3e9718f8d611bffe7d7fa17be91e3e79faec0e86d888c1a0776f001a4c1477a5"} Nov 24 07:09:12 crc kubenswrapper[4809]: E1124 07:09:12.841937 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4xjsj" podUID="39fc255b-7ac9-40b4-98e0-74d1f76c2696" Nov 24 07:09:12 crc kubenswrapper[4809]: I1124 07:09:12.846880 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-4vrfb" event={"ID":"5b90cc28-7c58-4826-b849-3456bb654f0d","Type":"ContainerStarted","Data":"8dbe166fd4eb5d9e97d68ec810f818e9363e60ae68476f3bcf14d3f6f05fc4af"} Nov 24 07:09:12 crc kubenswrapper[4809]: I1124 07:09:12.847048 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-4vrfb" event={"ID":"5b90cc28-7c58-4826-b849-3456bb654f0d","Type":"ContainerStarted","Data":"948766595ea70efe50bea6450d549a396901b495b9b7ba0c0ca02cde51fb189a"} Nov 24 07:09:12 crc kubenswrapper[4809]: I1124 07:09:12.854353 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-px9vl" event={"ID":"02a1aa7c-c3f7-4dcb-89ad-8d34f46ff4a1","Type":"ContainerStarted","Data":"bf81f2f4045f227f284e56c26c9b0dcfc00f577628326da061bed19e37d4af12"} Nov 24 07:09:12 crc kubenswrapper[4809]: I1124 07:09:12.856535 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8fcb84d5c-xqd67" event={"ID":"5b38e6fb-1aef-4ecd-80ca-4665a9d3430a","Type":"ContainerStarted","Data":"84cf97fc6b683ae3eddc87f08dc9974c20eca119366291080a1ea1eada737310"} Nov 24 07:09:12 crc kubenswrapper[4809]: I1124 07:09:12.856559 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8fcb84d5c-xqd67" event={"ID":"5b38e6fb-1aef-4ecd-80ca-4665a9d3430a","Type":"ContainerStarted","Data":"a4ec65f027b7b37c6c802c1c39d393774f84eccd4fa8ab34ce83af528424f08c"} Nov 24 07:09:12 crc kubenswrapper[4809]: I1124 07:09:12.856570 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8fcb84d5c-xqd67" event={"ID":"5b38e6fb-1aef-4ecd-80ca-4665a9d3430a","Type":"ContainerStarted","Data":"9f23aad8fb61114cc4507d41faf2a8a555d3fd2e8df0690ad0706c894a81bd74"} Nov 24 07:09:12 crc kubenswrapper[4809]: I1124 07:09:12.857161 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-8fcb84d5c-xqd67" Nov 24 07:09:12 crc kubenswrapper[4809]: I1124 07:09:12.859295 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-tk2hk" event={"ID":"fd9ddd03-bc95-43e1-93ef-97ad5ca589b3","Type":"ContainerStarted","Data":"1705fb7e5254ce5d3f87c3db8d4ad12cd93c7245dbe1a4e5249b16151842218a"} Nov 24 07:09:12 crc kubenswrapper[4809]: I1124 07:09:12.859323 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-tk2hk" event={"ID":"fd9ddd03-bc95-43e1-93ef-97ad5ca589b3","Type":"ContainerStarted","Data":"7ced5600ff67550fe7def3e72159eea4508e4c692d1977a87058be5de34a72c9"} Nov 24 07:09:12 crc kubenswrapper[4809]: I1124 07:09:12.860510 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-wbzcn" event={"ID":"cb240075-d846-4751-b026-6fa452a565a5","Type":"ContainerStarted","Data":"d5d9ed0b51752949a84b5b2be41fd3e01348269d0a8e9a6779b519520ad544ab"} Nov 24 07:09:12 crc kubenswrapper[4809]: I1124 07:09:12.860542 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-wbzcn" event={"ID":"cb240075-d846-4751-b026-6fa452a565a5","Type":"ContainerStarted","Data":"a07ec5d31b5882da3a34737e5074a9c8e28cd855d36b12fc0ffa08db74c8a1f3"} Nov 24 07:09:12 crc kubenswrapper[4809]: E1124 07:09:12.861412 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894\\\"\"" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-4vrfb" podUID="5b90cc28-7c58-4826-b849-3456bb654f0d" Nov 24 07:09:12 crc kubenswrapper[4809]: E1124 07:09:12.862402 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\"" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-tk2hk" podUID="fd9ddd03-bc95-43e1-93ef-97ad5ca589b3" Nov 24 07:09:12 crc kubenswrapper[4809]: E1124 07:09:12.863275 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-wbzcn" podUID="cb240075-d846-4751-b026-6fa452a565a5" Nov 24 07:09:13 crc kubenswrapper[4809]: I1124 07:09:13.016844 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-8fcb84d5c-xqd67" podStartSLOduration=3.016825481 podStartE2EDuration="3.016825481s" podCreationTimestamp="2025-11-24 07:09:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:09:12.983410409 +0000 UTC m=+852.884002114" watchObservedRunningTime="2025-11-24 07:09:13.016825481 +0000 UTC m=+852.917417176" Nov 24 07:09:13 crc kubenswrapper[4809]: E1124 07:09:13.877348 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-59brn" podUID="9eeb9814-10f8-478d-888f-5cf058c99c4b" Nov 24 07:09:13 crc kubenswrapper[4809]: E1124 07:09:13.877457 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\"" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-tk2hk" podUID="fd9ddd03-bc95-43e1-93ef-97ad5ca589b3" Nov 24 07:09:13 crc kubenswrapper[4809]: E1124 07:09:13.881109 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-7vg75" podUID="2e599303-b44b-4ce5-9ee9-1e1c13ca856f" Nov 24 07:09:13 crc kubenswrapper[4809]: E1124 07:09:13.889068 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894\\\"\"" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-4vrfb" podUID="5b90cc28-7c58-4826-b849-3456bb654f0d" Nov 24 07:09:13 crc kubenswrapper[4809]: E1124 07:09:13.889155 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4xjsj" podUID="39fc255b-7ac9-40b4-98e0-74d1f76c2696" Nov 24 07:09:13 crc kubenswrapper[4809]: E1124 07:09:13.905360 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-wbzcn" podUID="cb240075-d846-4751-b026-6fa452a565a5" Nov 24 07:09:21 crc kubenswrapper[4809]: I1124 07:09:21.321064 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-8fcb84d5c-xqd67" Nov 24 07:09:25 crc kubenswrapper[4809]: E1124 07:09:25.001912 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Nov 24 07:09:25 crc kubenswrapper[4809]: E1124 07:09:25.002616 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wzhnm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-px9vl_openstack-operators(02a1aa7c-c3f7-4dcb-89ad-8d34f46ff4a1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 07:09:25 crc kubenswrapper[4809]: E1124 07:09:25.003827 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-px9vl" podUID="02a1aa7c-c3f7-4dcb-89ad-8d34f46ff4a1" Nov 24 07:09:25 crc kubenswrapper[4809]: E1124 07:09:25.383920 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f" Nov 24 07:09:25 crc kubenswrapper[4809]: E1124 07:09:25.384124 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2hhkz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-8c6448b9f-twjz4_openstack-operators(df289b0d-b465-49d2-b1ee-4f0faf87b079): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 07:09:25 crc kubenswrapper[4809]: I1124 07:09:25.955673 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-rgkd9" event={"ID":"e94d989e-7aee-43d9-87c0-712355ca193f","Type":"ContainerStarted","Data":"3f64482c82a5c786eeb8d0245625ab73ad16b711146a950996992c7af919cf7a"} Nov 24 07:09:25 crc kubenswrapper[4809]: E1124 07:09:25.956878 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-px9vl" podUID="02a1aa7c-c3f7-4dcb-89ad-8d34f46ff4a1" Nov 24 07:09:26 crc kubenswrapper[4809]: I1124 07:09:26.922457 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tnwx9"] Nov 24 07:09:26 crc kubenswrapper[4809]: I1124 07:09:26.924199 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tnwx9" Nov 24 07:09:26 crc kubenswrapper[4809]: I1124 07:09:26.933311 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tnwx9"] Nov 24 07:09:27 crc kubenswrapper[4809]: I1124 07:09:27.063427 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpj9r\" (UniqueName: \"kubernetes.io/projected/da3c2694-88ce-4440-a0de-f63475eff3cf-kube-api-access-hpj9r\") pod \"redhat-operators-tnwx9\" (UID: \"da3c2694-88ce-4440-a0de-f63475eff3cf\") " pod="openshift-marketplace/redhat-operators-tnwx9" Nov 24 07:09:27 crc kubenswrapper[4809]: I1124 07:09:27.063531 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da3c2694-88ce-4440-a0de-f63475eff3cf-catalog-content\") pod \"redhat-operators-tnwx9\" (UID: \"da3c2694-88ce-4440-a0de-f63475eff3cf\") " pod="openshift-marketplace/redhat-operators-tnwx9" Nov 24 07:09:27 crc kubenswrapper[4809]: I1124 07:09:27.063563 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da3c2694-88ce-4440-a0de-f63475eff3cf-utilities\") pod \"redhat-operators-tnwx9\" (UID: \"da3c2694-88ce-4440-a0de-f63475eff3cf\") " pod="openshift-marketplace/redhat-operators-tnwx9" Nov 24 07:09:27 crc kubenswrapper[4809]: I1124 07:09:27.165296 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpj9r\" (UniqueName: \"kubernetes.io/projected/da3c2694-88ce-4440-a0de-f63475eff3cf-kube-api-access-hpj9r\") pod \"redhat-operators-tnwx9\" (UID: \"da3c2694-88ce-4440-a0de-f63475eff3cf\") " pod="openshift-marketplace/redhat-operators-tnwx9" Nov 24 07:09:27 crc kubenswrapper[4809]: I1124 07:09:27.165404 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da3c2694-88ce-4440-a0de-f63475eff3cf-catalog-content\") pod \"redhat-operators-tnwx9\" (UID: \"da3c2694-88ce-4440-a0de-f63475eff3cf\") " pod="openshift-marketplace/redhat-operators-tnwx9" Nov 24 07:09:27 crc kubenswrapper[4809]: I1124 07:09:27.165442 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da3c2694-88ce-4440-a0de-f63475eff3cf-utilities\") pod \"redhat-operators-tnwx9\" (UID: \"da3c2694-88ce-4440-a0de-f63475eff3cf\") " pod="openshift-marketplace/redhat-operators-tnwx9" Nov 24 07:09:27 crc kubenswrapper[4809]: I1124 07:09:27.165951 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da3c2694-88ce-4440-a0de-f63475eff3cf-utilities\") pod \"redhat-operators-tnwx9\" (UID: \"da3c2694-88ce-4440-a0de-f63475eff3cf\") " pod="openshift-marketplace/redhat-operators-tnwx9" Nov 24 07:09:27 crc kubenswrapper[4809]: I1124 07:09:27.166146 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da3c2694-88ce-4440-a0de-f63475eff3cf-catalog-content\") pod \"redhat-operators-tnwx9\" (UID: \"da3c2694-88ce-4440-a0de-f63475eff3cf\") " pod="openshift-marketplace/redhat-operators-tnwx9" Nov 24 07:09:27 crc kubenswrapper[4809]: I1124 07:09:27.195613 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpj9r\" (UniqueName: \"kubernetes.io/projected/da3c2694-88ce-4440-a0de-f63475eff3cf-kube-api-access-hpj9r\") pod \"redhat-operators-tnwx9\" (UID: \"da3c2694-88ce-4440-a0de-f63475eff3cf\") " pod="openshift-marketplace/redhat-operators-tnwx9" Nov 24 07:09:27 crc kubenswrapper[4809]: I1124 07:09:27.242894 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tnwx9" Nov 24 07:09:27 crc kubenswrapper[4809]: I1124 07:09:27.676456 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tnwx9"] Nov 24 07:09:27 crc kubenswrapper[4809]: I1124 07:09:27.970646 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tnwx9" event={"ID":"da3c2694-88ce-4440-a0de-f63475eff3cf","Type":"ContainerStarted","Data":"1214be828e40a40b5c1fca3e9108314953c2c1cebe843c49c5c1afd694656ff3"} Nov 24 07:09:28 crc kubenswrapper[4809]: I1124 07:09:28.978208 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-8nzmk" event={"ID":"0a1424d6-9ecf-4a06-b174-15d13f61b916","Type":"ContainerStarted","Data":"799ea7b0a1bab7688570cbc9e9bec9c3b80fc4608b030bf2a1b37687ef36bce5"} Nov 24 07:09:29 crc kubenswrapper[4809]: I1124 07:09:29.985596 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-xf4fh" event={"ID":"4d4aa35c-faa7-4b93-aecd-ee4645f11916","Type":"ContainerStarted","Data":"5989ca7031f5fa74338fa33d5467964cb9338324fe793e89c6bd2bf535604ec4"} Nov 24 07:09:29 crc kubenswrapper[4809]: I1124 07:09:29.987126 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-49dpw" event={"ID":"e525562e-b6b2-415f-b6e7-1c0bfde8a56f","Type":"ContainerStarted","Data":"8192721e9f9d700b4f7bd9d300d5185fbcf68a7bd416909b3bb019ebf3606026"} Nov 24 07:09:29 crc kubenswrapper[4809]: I1124 07:09:29.988633 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-twjz4" event={"ID":"df289b0d-b465-49d2-b1ee-4f0faf87b079","Type":"ContainerStarted","Data":"66f77d3eb82eb7d61463651e0fdb596c4a846dcbc9aa380a0bd9cbe5bd4a98aa"} Nov 24 07:09:29 crc kubenswrapper[4809]: I1124 07:09:29.990094 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-wh5fq" event={"ID":"b0a19ac9-d266-4416-8590-3c8de4dadc7f","Type":"ContainerStarted","Data":"713a04124fa5202446fa32a2638a6daba19ab1f1f96ceedc648abddea8b23ed3"} Nov 24 07:09:29 crc kubenswrapper[4809]: I1124 07:09:29.995574 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-plhlb" event={"ID":"200f6896-3100-46ea-94f4-69637762a259","Type":"ContainerStarted","Data":"d98e2a8a5b3fb57b117191dca36f296385e4b7d6ac255fe92a62c11703618513"} Nov 24 07:09:29 crc kubenswrapper[4809]: I1124 07:09:29.997164 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-bz8c8" event={"ID":"a2b969b3-07c7-4bf3-ab18-5317d8f0a372","Type":"ContainerStarted","Data":"445e56e984c68ffd705862a6c6dfcbb0544b65bc1612b5812ddbecede1d7954f"} Nov 24 07:09:29 crc kubenswrapper[4809]: I1124 07:09:29.998716 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-84nbm" event={"ID":"bfe0f53e-685b-4cef-b763-90399c5fdaac","Type":"ContainerStarted","Data":"9edcba0ce72491e16cc8d9525670f3f7da5e4fae2caf4c561fc400a342a045ea"} Nov 24 07:09:30 crc kubenswrapper[4809]: I1124 07:09:30.000158 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jfjqt" event={"ID":"5f4f9ac0-6714-4f4a-a7fb-8d65b19c919c","Type":"ContainerStarted","Data":"c3ac36c0b4c8ec1200de6a8495b55d667a05ac40074311d02bd8e2176389ad0b"} Nov 24 07:09:30 crc kubenswrapper[4809]: I1124 07:09:30.001418 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-4ddch" event={"ID":"d78cc4fe-bab9-4521-a1d2-ed6efd341789","Type":"ContainerStarted","Data":"4580894ca19ff7ff594106c8102dba130a133e7d30e9ae14a4f8a3fcc5eb01ff"} Nov 24 07:09:30 crc kubenswrapper[4809]: I1124 07:09:30.004672 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rtxsf" event={"ID":"483b5843-7751-465a-9412-dbb15ef5145f","Type":"ContainerStarted","Data":"024e64466da0f693ea3c185320f691cfb48eec411ba673fab262c2e8b1961d43"} Nov 24 07:09:30 crc kubenswrapper[4809]: I1124 07:09:30.006878 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-rgkd9" event={"ID":"e94d989e-7aee-43d9-87c0-712355ca193f","Type":"ContainerStarted","Data":"7f57b67256e4a3fe1e40943a051795221c93e385257444e442dcd1df21755b53"} Nov 24 07:09:30 crc kubenswrapper[4809]: I1124 07:09:30.007842 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-rgkd9" Nov 24 07:09:30 crc kubenswrapper[4809]: I1124 07:09:30.009570 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-rgkd9" Nov 24 07:09:30 crc kubenswrapper[4809]: I1124 07:09:30.009859 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-rp7r4" event={"ID":"2f364be5-58b1-4d57-97f5-65a13d63e34e","Type":"ContainerStarted","Data":"52f1564a30aad5ae342e57e2c2a38ea665e94779f6dd6c232a4fd8e76ce673dc"} Nov 24 07:09:30 crc kubenswrapper[4809]: I1124 07:09:30.011982 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-mfgld" event={"ID":"6269a5dc-fd2a-4fdd-9b18-4e505242efca","Type":"ContainerStarted","Data":"6a10897b1c8fdd3363145c1cf37607fd1b75caacb9c6f5384d9651051934f2a7"} Nov 24 07:09:30 crc kubenswrapper[4809]: I1124 07:09:30.013382 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-6k9xr" event={"ID":"bfcf902b-83f0-48f9-ad6b-eb7963f43af4","Type":"ContainerStarted","Data":"c1da86337d08178ab94c9fc8ac4cf87f5bbd2634f99adb5b151dad09ea61829b"} Nov 24 07:09:30 crc kubenswrapper[4809]: I1124 07:09:30.043737 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-rgkd9" podStartSLOduration=7.216151271 podStartE2EDuration="21.043716208s" podCreationTimestamp="2025-11-24 07:09:09 +0000 UTC" firstStartedPulling="2025-11-24 07:09:11.535190244 +0000 UTC m=+851.435781949" lastFinishedPulling="2025-11-24 07:09:25.362755171 +0000 UTC m=+865.263346886" observedRunningTime="2025-11-24 07:09:30.021525597 +0000 UTC m=+869.922117302" watchObservedRunningTime="2025-11-24 07:09:30.043716208 +0000 UTC m=+869.944307913" Nov 24 07:09:30 crc kubenswrapper[4809]: E1124 07:09:30.125342 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-twjz4" podUID="df289b0d-b465-49d2-b1ee-4f0faf87b079" Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.031907 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-xf4fh" event={"ID":"4d4aa35c-faa7-4b93-aecd-ee4645f11916","Type":"ContainerStarted","Data":"c38a7ae77fd0eb61dd7baa21ed935d90c23bc1897f4dd053bb77743ad4b14c27"} Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.032480 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-xf4fh" Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.046571 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-plhlb" event={"ID":"200f6896-3100-46ea-94f4-69637762a259","Type":"ContainerStarted","Data":"ec888724271176b23a3c15e1146552c6a3208f9ffb4cc488aaeb909000c2fe92"} Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.051941 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-84nbm" event={"ID":"bfe0f53e-685b-4cef-b763-90399c5fdaac","Type":"ContainerStarted","Data":"af692a95e108ae40a11eb48cfdd7f2cb0ac97f05df4aadb026527c4c0db678b0"} Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.052640 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-84nbm" Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.055337 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-mfgld" event={"ID":"6269a5dc-fd2a-4fdd-9b18-4e505242efca","Type":"ContainerStarted","Data":"24b1eab87fdd9a36b566649d98bd39a7b3b216372b1525de3f65c1365fb3ba96"} Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.055785 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-d656998f4-mfgld" Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.060933 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-6k9xr" event={"ID":"bfcf902b-83f0-48f9-ad6b-eb7963f43af4","Type":"ContainerStarted","Data":"f04d5361b147da16c41dfe49ffa195285773ce030792088b4590f0a07a671d59"} Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.061349 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-xf4fh" podStartSLOduration=8.320174003 podStartE2EDuration="22.061331753s" podCreationTimestamp="2025-11-24 07:09:09 +0000 UTC" firstStartedPulling="2025-11-24 07:09:11.623709727 +0000 UTC m=+851.524301432" lastFinishedPulling="2025-11-24 07:09:25.364867477 +0000 UTC m=+865.265459182" observedRunningTime="2025-11-24 07:09:31.057602889 +0000 UTC m=+870.958194594" watchObservedRunningTime="2025-11-24 07:09:31.061331753 +0000 UTC m=+870.961923458" Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.061527 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-6k9xr" Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.065755 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jfjqt" event={"ID":"5f4f9ac0-6714-4f4a-a7fb-8d65b19c919c","Type":"ContainerStarted","Data":"090775c19be2bb35c6f0c4b8d4a91563b0895fd16c4c77e5eef88414b8db9e99"} Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.066144 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jfjqt" Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.067699 4809 generic.go:334] "Generic (PLEG): container finished" podID="da3c2694-88ce-4440-a0de-f63475eff3cf" containerID="c0f5bd32a98731969fcb35d06511594d307641369b6619ce0b13199ba97b6362" exitCode=0 Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.067778 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tnwx9" event={"ID":"da3c2694-88ce-4440-a0de-f63475eff3cf","Type":"ContainerDied","Data":"c0f5bd32a98731969fcb35d06511594d307641369b6619ce0b13199ba97b6362"} Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.069874 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-rp7r4" event={"ID":"2f364be5-58b1-4d57-97f5-65a13d63e34e","Type":"ContainerStarted","Data":"b9019e70d0f6baf00d2577d8319b8db55007dd6cf75286a263beb5453a5e59f8"} Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.070006 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-rp7r4" Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.072390 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-4ddch" event={"ID":"d78cc4fe-bab9-4521-a1d2-ed6efd341789","Type":"ContainerStarted","Data":"bef158143f12c56eeedf468e74ccf8e90d34ff8e272798da35494a8152ac0a7b"} Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.072726 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-4ddch" Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.079268 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rtxsf" event={"ID":"483b5843-7751-465a-9412-dbb15ef5145f","Type":"ContainerStarted","Data":"185dd1ec9596bda363c39ca4d1c187eae299894cf023e001ae8372f89bfa79cd"} Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.079900 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rtxsf" Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.084498 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-wh5fq" event={"ID":"b0a19ac9-d266-4416-8590-3c8de4dadc7f","Type":"ContainerStarted","Data":"99289a41b4725aa308e08f6f797bb2a147bf69b5530c5b842144d960ffd410b3"} Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.084956 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-wh5fq" Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.087680 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-8nzmk" event={"ID":"0a1424d6-9ecf-4a06-b174-15d13f61b916","Type":"ContainerStarted","Data":"6845f6d3f42161f6b9040a1b6cd1d2b5fd10dd41a0ef3c99c179aaab5ddbc5de"} Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.088059 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7969689c84-8nzmk" Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.090353 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-bz8c8" event={"ID":"a2b969b3-07c7-4bf3-ab18-5317d8f0a372","Type":"ContainerStarted","Data":"3752b57df187597dc2143759774c6802795ddc3bd35f6fafb336c0b152df9be9"} Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.090376 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-58f887965d-bz8c8" Nov 24 07:09:31 crc kubenswrapper[4809]: E1124 07:09:31.091221 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-twjz4" podUID="df289b0d-b465-49d2-b1ee-4f0faf87b079" Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.104318 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-84nbm" podStartSLOduration=8.27183346 podStartE2EDuration="22.104303696s" podCreationTimestamp="2025-11-24 07:09:09 +0000 UTC" firstStartedPulling="2025-11-24 07:09:11.564049354 +0000 UTC m=+851.464641059" lastFinishedPulling="2025-11-24 07:09:25.39651958 +0000 UTC m=+865.297111295" observedRunningTime="2025-11-24 07:09:31.080115289 +0000 UTC m=+870.980706994" watchObservedRunningTime="2025-11-24 07:09:31.104303696 +0000 UTC m=+871.004895401" Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.131536 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-d656998f4-mfgld" podStartSLOduration=8.400666318 podStartE2EDuration="22.13151791s" podCreationTimestamp="2025-11-24 07:09:09 +0000 UTC" firstStartedPulling="2025-11-24 07:09:11.632217556 +0000 UTC m=+851.532809261" lastFinishedPulling="2025-11-24 07:09:25.363069148 +0000 UTC m=+865.263660853" observedRunningTime="2025-11-24 07:09:31.106939124 +0000 UTC m=+871.007530829" watchObservedRunningTime="2025-11-24 07:09:31.13151791 +0000 UTC m=+871.032109615" Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.148285 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-plhlb" podStartSLOduration=8.30925594 podStartE2EDuration="22.148268901s" podCreationTimestamp="2025-11-24 07:09:09 +0000 UTC" firstStartedPulling="2025-11-24 07:09:11.536746878 +0000 UTC m=+851.437338583" lastFinishedPulling="2025-11-24 07:09:25.375759829 +0000 UTC m=+865.276351544" observedRunningTime="2025-11-24 07:09:31.127617953 +0000 UTC m=+871.028209658" watchObservedRunningTime="2025-11-24 07:09:31.148268901 +0000 UTC m=+871.048860606" Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.167238 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7969689c84-8nzmk" podStartSLOduration=7.435104 podStartE2EDuration="22.167220982s" podCreationTimestamp="2025-11-24 07:09:09 +0000 UTC" firstStartedPulling="2025-11-24 07:09:10.633326688 +0000 UTC m=+850.533918393" lastFinishedPulling="2025-11-24 07:09:25.36544367 +0000 UTC m=+865.266035375" observedRunningTime="2025-11-24 07:09:31.166073166 +0000 UTC m=+871.066664871" watchObservedRunningTime="2025-11-24 07:09:31.167220982 +0000 UTC m=+871.067812677" Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.187520 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-6k9xr" podStartSLOduration=7.979736531 podStartE2EDuration="22.187505842s" podCreationTimestamp="2025-11-24 07:09:09 +0000 UTC" firstStartedPulling="2025-11-24 07:09:11.15815662 +0000 UTC m=+851.058748325" lastFinishedPulling="2025-11-24 07:09:25.365925931 +0000 UTC m=+865.266517636" observedRunningTime="2025-11-24 07:09:31.186307585 +0000 UTC m=+871.086899290" watchObservedRunningTime="2025-11-24 07:09:31.187505842 +0000 UTC m=+871.088097547" Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.211770 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-rp7r4" podStartSLOduration=7.9085706810000005 podStartE2EDuration="22.211755439s" podCreationTimestamp="2025-11-24 07:09:09 +0000 UTC" firstStartedPulling="2025-11-24 07:09:11.071271673 +0000 UTC m=+850.971863378" lastFinishedPulling="2025-11-24 07:09:25.374456421 +0000 UTC m=+865.275048136" observedRunningTime="2025-11-24 07:09:31.207789051 +0000 UTC m=+871.108380756" watchObservedRunningTime="2025-11-24 07:09:31.211755439 +0000 UTC m=+871.112347144" Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.228838 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rtxsf" podStartSLOduration=8.464954613 podStartE2EDuration="22.228822168s" podCreationTimestamp="2025-11-24 07:09:09 +0000 UTC" firstStartedPulling="2025-11-24 07:09:11.607927197 +0000 UTC m=+851.508518902" lastFinishedPulling="2025-11-24 07:09:25.371794752 +0000 UTC m=+865.272386457" observedRunningTime="2025-11-24 07:09:31.226913876 +0000 UTC m=+871.127505581" watchObservedRunningTime="2025-11-24 07:09:31.228822168 +0000 UTC m=+871.129413873" Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.248287 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-4ddch" podStartSLOduration=7.685875891 podStartE2EDuration="22.248272759s" podCreationTimestamp="2025-11-24 07:09:09 +0000 UTC" firstStartedPulling="2025-11-24 07:09:10.803612295 +0000 UTC m=+850.704204000" lastFinishedPulling="2025-11-24 07:09:25.366009153 +0000 UTC m=+865.266600868" observedRunningTime="2025-11-24 07:09:31.245601 +0000 UTC m=+871.146192725" watchObservedRunningTime="2025-11-24 07:09:31.248272759 +0000 UTC m=+871.148864464" Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.287241 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jfjqt" podStartSLOduration=7.890283195 podStartE2EDuration="22.287227303s" podCreationTimestamp="2025-11-24 07:09:09 +0000 UTC" firstStartedPulling="2025-11-24 07:09:10.977670886 +0000 UTC m=+850.878262591" lastFinishedPulling="2025-11-24 07:09:25.374614984 +0000 UTC m=+865.275206699" observedRunningTime="2025-11-24 07:09:31.284508463 +0000 UTC m=+871.185100168" watchObservedRunningTime="2025-11-24 07:09:31.287227303 +0000 UTC m=+871.187818998" Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.324536 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-58f887965d-bz8c8" podStartSLOduration=8.472051621 podStartE2EDuration="22.324521631s" podCreationTimestamp="2025-11-24 07:09:09 +0000 UTC" firstStartedPulling="2025-11-24 07:09:11.539866547 +0000 UTC m=+851.440458252" lastFinishedPulling="2025-11-24 07:09:25.392336547 +0000 UTC m=+865.292928262" observedRunningTime="2025-11-24 07:09:31.323236502 +0000 UTC m=+871.223828217" watchObservedRunningTime="2025-11-24 07:09:31.324521631 +0000 UTC m=+871.225113326" Nov 24 07:09:31 crc kubenswrapper[4809]: I1124 07:09:31.325631 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-wh5fq" podStartSLOduration=7.916252042 podStartE2EDuration="22.325624375s" podCreationTimestamp="2025-11-24 07:09:09 +0000 UTC" firstStartedPulling="2025-11-24 07:09:10.965289362 +0000 UTC m=+850.865881067" lastFinishedPulling="2025-11-24 07:09:25.374661695 +0000 UTC m=+865.275253400" observedRunningTime="2025-11-24 07:09:31.306678255 +0000 UTC m=+871.207269960" watchObservedRunningTime="2025-11-24 07:09:31.325624375 +0000 UTC m=+871.226216070" Nov 24 07:09:32 crc kubenswrapper[4809]: I1124 07:09:32.110325 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-7vg75" event={"ID":"2e599303-b44b-4ce5-9ee9-1e1c13ca856f","Type":"ContainerStarted","Data":"08bcdd7aa7e841b82a3e3053cced2ee2af9bc8da26bdc93b58f3a5b1945c3e43"} Nov 24 07:09:32 crc kubenswrapper[4809]: I1124 07:09:32.110915 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-7vg75" Nov 24 07:09:32 crc kubenswrapper[4809]: I1124 07:09:32.113527 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-49dpw" event={"ID":"e525562e-b6b2-415f-b6e7-1c0bfde8a56f","Type":"ContainerStarted","Data":"be235c4e091fcf08ea42407b3d6174a11658ac9aef75459d0f9bdf9ba7dc59e2"} Nov 24 07:09:32 crc kubenswrapper[4809]: I1124 07:09:32.114231 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-plhlb" Nov 24 07:09:32 crc kubenswrapper[4809]: I1124 07:09:32.143019 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-7vg75" podStartSLOduration=3.384242919 podStartE2EDuration="23.143001748s" podCreationTimestamp="2025-11-24 07:09:09 +0000 UTC" firstStartedPulling="2025-11-24 07:09:11.903183897 +0000 UTC m=+851.803775602" lastFinishedPulling="2025-11-24 07:09:31.661942726 +0000 UTC m=+871.562534431" observedRunningTime="2025-11-24 07:09:32.141151706 +0000 UTC m=+872.041743421" watchObservedRunningTime="2025-11-24 07:09:32.143001748 +0000 UTC m=+872.043593453" Nov 24 07:09:32 crc kubenswrapper[4809]: I1124 07:09:32.160032 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-49dpw" podStartSLOduration=8.401073338 podStartE2EDuration="23.160009545s" podCreationTimestamp="2025-11-24 07:09:09 +0000 UTC" firstStartedPulling="2025-11-24 07:09:10.605404619 +0000 UTC m=+850.505996324" lastFinishedPulling="2025-11-24 07:09:25.364340826 +0000 UTC m=+865.264932531" observedRunningTime="2025-11-24 07:09:32.156219141 +0000 UTC m=+872.056810846" watchObservedRunningTime="2025-11-24 07:09:32.160009545 +0000 UTC m=+872.060601250" Nov 24 07:09:33 crc kubenswrapper[4809]: I1124 07:09:33.120042 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-49dpw" Nov 24 07:09:33 crc kubenswrapper[4809]: I1124 07:09:33.123172 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7969689c84-8nzmk" Nov 24 07:09:34 crc kubenswrapper[4809]: I1124 07:09:34.134868 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-49dpw" Nov 24 07:09:35 crc kubenswrapper[4809]: I1124 07:09:35.139237 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-tk2hk" event={"ID":"fd9ddd03-bc95-43e1-93ef-97ad5ca589b3","Type":"ContainerStarted","Data":"443be5a8a8d990f93fdf2f9bacf38d8f5d25b89e87e1ce69fafba2862bad0fa1"} Nov 24 07:09:35 crc kubenswrapper[4809]: I1124 07:09:35.139459 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-tk2hk" Nov 24 07:09:35 crc kubenswrapper[4809]: I1124 07:09:35.141332 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-wbzcn" event={"ID":"cb240075-d846-4751-b026-6fa452a565a5","Type":"ContainerStarted","Data":"287171520fe259de40ed57b34fce70bd9eb19161bdc50be0aa245db674307134"} Nov 24 07:09:35 crc kubenswrapper[4809]: I1124 07:09:35.141540 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-b4c496f69-wbzcn" Nov 24 07:09:35 crc kubenswrapper[4809]: I1124 07:09:35.143128 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tnwx9" event={"ID":"da3c2694-88ce-4440-a0de-f63475eff3cf","Type":"ContainerStarted","Data":"bc284e794dd87a9d1d7c1f50c5ab0bff28956537b6ac2de3983e9078149e7d0a"} Nov 24 07:09:35 crc kubenswrapper[4809]: I1124 07:09:35.145061 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-59brn" event={"ID":"9eeb9814-10f8-478d-888f-5cf058c99c4b","Type":"ContainerStarted","Data":"f4f8e9dabb50395a829e68f5f1da70f66f8891a8ebbfcaa001bc43d58e33e7c3"} Nov 24 07:09:35 crc kubenswrapper[4809]: I1124 07:09:35.145256 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-59brn" Nov 24 07:09:35 crc kubenswrapper[4809]: I1124 07:09:35.147271 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4xjsj" event={"ID":"39fc255b-7ac9-40b4-98e0-74d1f76c2696","Type":"ContainerStarted","Data":"2cb346436d2e069a48111801484b2e84bcc38b660981f0ed8f0ec46c8388c6c1"} Nov 24 07:09:35 crc kubenswrapper[4809]: I1124 07:09:35.147489 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4xjsj" Nov 24 07:09:35 crc kubenswrapper[4809]: I1124 07:09:35.149619 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-4vrfb" event={"ID":"5b90cc28-7c58-4826-b849-3456bb654f0d","Type":"ContainerStarted","Data":"9602b5c3e58515472bc9ba0d74bd9beef2b90f4418d90de881547116f2015119"} Nov 24 07:09:35 crc kubenswrapper[4809]: I1124 07:09:35.149931 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-4vrfb" Nov 24 07:09:35 crc kubenswrapper[4809]: I1124 07:09:35.190512 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-59brn" podStartSLOduration=3.34643916 podStartE2EDuration="26.19049258s" podCreationTimestamp="2025-11-24 07:09:09 +0000 UTC" firstStartedPulling="2025-11-24 07:09:11.828439489 +0000 UTC m=+851.729031194" lastFinishedPulling="2025-11-24 07:09:34.672492869 +0000 UTC m=+874.573084614" observedRunningTime="2025-11-24 07:09:35.184713942 +0000 UTC m=+875.085305647" watchObservedRunningTime="2025-11-24 07:09:35.19049258 +0000 UTC m=+875.091084285" Nov 24 07:09:35 crc kubenswrapper[4809]: I1124 07:09:35.190901 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-tk2hk" podStartSLOduration=3.361848162 podStartE2EDuration="26.190891849s" podCreationTimestamp="2025-11-24 07:09:09 +0000 UTC" firstStartedPulling="2025-11-24 07:09:11.82848463 +0000 UTC m=+851.729076335" lastFinishedPulling="2025-11-24 07:09:34.657528317 +0000 UTC m=+874.558120022" observedRunningTime="2025-11-24 07:09:35.165409014 +0000 UTC m=+875.066000719" watchObservedRunningTime="2025-11-24 07:09:35.190891849 +0000 UTC m=+875.091483554" Nov 24 07:09:35 crc kubenswrapper[4809]: I1124 07:09:35.210594 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4xjsj" podStartSLOduration=3.456162063 podStartE2EDuration="26.210574865s" podCreationTimestamp="2025-11-24 07:09:09 +0000 UTC" firstStartedPulling="2025-11-24 07:09:11.903288399 +0000 UTC m=+851.803880104" lastFinishedPulling="2025-11-24 07:09:34.657701211 +0000 UTC m=+874.558292906" observedRunningTime="2025-11-24 07:09:35.207930687 +0000 UTC m=+875.108522392" watchObservedRunningTime="2025-11-24 07:09:35.210574865 +0000 UTC m=+875.111166580" Nov 24 07:09:35 crc kubenswrapper[4809]: I1124 07:09:35.227609 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-b4c496f69-wbzcn" podStartSLOduration=2.412295371 podStartE2EDuration="25.227594153s" podCreationTimestamp="2025-11-24 07:09:10 +0000 UTC" firstStartedPulling="2025-11-24 07:09:11.837737546 +0000 UTC m=+851.738329251" lastFinishedPulling="2025-11-24 07:09:34.653036328 +0000 UTC m=+874.553628033" observedRunningTime="2025-11-24 07:09:35.224697719 +0000 UTC m=+875.125289424" watchObservedRunningTime="2025-11-24 07:09:35.227594153 +0000 UTC m=+875.128185858" Nov 24 07:09:35 crc kubenswrapper[4809]: I1124 07:09:35.247438 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-4vrfb" podStartSLOduration=3.396508381 podStartE2EDuration="26.247424123s" podCreationTimestamp="2025-11-24 07:09:09 +0000 UTC" firstStartedPulling="2025-11-24 07:09:11.8203825 +0000 UTC m=+851.720974195" lastFinishedPulling="2025-11-24 07:09:34.671298232 +0000 UTC m=+874.571889937" observedRunningTime="2025-11-24 07:09:35.243009824 +0000 UTC m=+875.143601529" watchObservedRunningTime="2025-11-24 07:09:35.247424123 +0000 UTC m=+875.148015828" Nov 24 07:09:36 crc kubenswrapper[4809]: I1124 07:09:36.158272 4809 generic.go:334] "Generic (PLEG): container finished" podID="da3c2694-88ce-4440-a0de-f63475eff3cf" containerID="bc284e794dd87a9d1d7c1f50c5ab0bff28956537b6ac2de3983e9078149e7d0a" exitCode=0 Nov 24 07:09:36 crc kubenswrapper[4809]: I1124 07:09:36.158354 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tnwx9" event={"ID":"da3c2694-88ce-4440-a0de-f63475eff3cf","Type":"ContainerDied","Data":"bc284e794dd87a9d1d7c1f50c5ab0bff28956537b6ac2de3983e9078149e7d0a"} Nov 24 07:09:37 crc kubenswrapper[4809]: I1124 07:09:37.169141 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tnwx9" event={"ID":"da3c2694-88ce-4440-a0de-f63475eff3cf","Type":"ContainerStarted","Data":"9e5167351e794b8ebfad543a0b0ff334e89bfde5e74d8feb40e8f51a5bbf6217"} Nov 24 07:09:37 crc kubenswrapper[4809]: I1124 07:09:37.195872 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tnwx9" podStartSLOduration=6.212280965 podStartE2EDuration="11.195855185s" podCreationTimestamp="2025-11-24 07:09:26 +0000 UTC" firstStartedPulling="2025-11-24 07:09:31.608873829 +0000 UTC m=+871.509465544" lastFinishedPulling="2025-11-24 07:09:36.592448019 +0000 UTC m=+876.493039764" observedRunningTime="2025-11-24 07:09:37.190912146 +0000 UTC m=+877.091503891" watchObservedRunningTime="2025-11-24 07:09:37.195855185 +0000 UTC m=+877.096446890" Nov 24 07:09:37 crc kubenswrapper[4809]: I1124 07:09:37.243017 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tnwx9" Nov 24 07:09:37 crc kubenswrapper[4809]: I1124 07:09:37.243077 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tnwx9" Nov 24 07:09:38 crc kubenswrapper[4809]: I1124 07:09:38.291286 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tnwx9" podUID="da3c2694-88ce-4440-a0de-f63475eff3cf" containerName="registry-server" probeResult="failure" output=< Nov 24 07:09:38 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Nov 24 07:09:38 crc kubenswrapper[4809]: > Nov 24 07:09:39 crc kubenswrapper[4809]: I1124 07:09:39.191866 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-px9vl" event={"ID":"02a1aa7c-c3f7-4dcb-89ad-8d34f46ff4a1","Type":"ContainerStarted","Data":"5c6ec8d10f7ba5701c9ce3ce21d05d9e61eb6c149893b5a2fa1a1a67e8d165b8"} Nov 24 07:09:39 crc kubenswrapper[4809]: I1124 07:09:39.210633 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-px9vl" podStartSLOduration=2.725015279 podStartE2EDuration="29.210618749s" podCreationTimestamp="2025-11-24 07:09:10 +0000 UTC" firstStartedPulling="2025-11-24 07:09:11.8190137 +0000 UTC m=+851.719605405" lastFinishedPulling="2025-11-24 07:09:38.30461714 +0000 UTC m=+878.205208875" observedRunningTime="2025-11-24 07:09:39.206242761 +0000 UTC m=+879.106834466" watchObservedRunningTime="2025-11-24 07:09:39.210618749 +0000 UTC m=+879.111210454" Nov 24 07:09:39 crc kubenswrapper[4809]: I1124 07:09:39.970414 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-4ddch" Nov 24 07:09:40 crc kubenswrapper[4809]: I1124 07:09:40.050130 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jfjqt" Nov 24 07:09:40 crc kubenswrapper[4809]: I1124 07:09:40.086522 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-wh5fq" Nov 24 07:09:40 crc kubenswrapper[4809]: I1124 07:09:40.138593 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-rp7r4" Nov 24 07:09:40 crc kubenswrapper[4809]: I1124 07:09:40.160818 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-6k9xr" Nov 24 07:09:40 crc kubenswrapper[4809]: I1124 07:09:40.168421 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-58f887965d-bz8c8" Nov 24 07:09:40 crc kubenswrapper[4809]: I1124 07:09:40.230593 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-plhlb" Nov 24 07:09:40 crc kubenswrapper[4809]: I1124 07:09:40.267275 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-xf4fh" Nov 24 07:09:40 crc kubenswrapper[4809]: I1124 07:09:40.377639 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-84nbm" Nov 24 07:09:40 crc kubenswrapper[4809]: I1124 07:09:40.424880 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-59brn" Nov 24 07:09:40 crc kubenswrapper[4809]: I1124 07:09:40.461860 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rtxsf" Nov 24 07:09:40 crc kubenswrapper[4809]: I1124 07:09:40.481394 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4xjsj" Nov 24 07:09:40 crc kubenswrapper[4809]: I1124 07:09:40.488505 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-d656998f4-mfgld" Nov 24 07:09:40 crc kubenswrapper[4809]: I1124 07:09:40.513491 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-b4c496f69-wbzcn" Nov 24 07:09:40 crc kubenswrapper[4809]: I1124 07:09:40.632447 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-tk2hk" Nov 24 07:09:40 crc kubenswrapper[4809]: I1124 07:09:40.633591 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-4vrfb" Nov 24 07:09:41 crc kubenswrapper[4809]: I1124 07:09:41.009859 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-7vg75" Nov 24 07:09:43 crc kubenswrapper[4809]: I1124 07:09:43.265386 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-twjz4" event={"ID":"df289b0d-b465-49d2-b1ee-4f0faf87b079","Type":"ContainerStarted","Data":"3378287a1fbfb32a356db5959e86dc2f608c9700b9953582a423c5c94db1af12"} Nov 24 07:09:43 crc kubenswrapper[4809]: I1124 07:09:43.266115 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-twjz4" Nov 24 07:09:43 crc kubenswrapper[4809]: I1124 07:09:43.283545 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-twjz4" podStartSLOduration=2.104489602 podStartE2EDuration="33.283528158s" podCreationTimestamp="2025-11-24 07:09:10 +0000 UTC" firstStartedPulling="2025-11-24 07:09:11.627828759 +0000 UTC m=+851.528420464" lastFinishedPulling="2025-11-24 07:09:42.806867305 +0000 UTC m=+882.707459020" observedRunningTime="2025-11-24 07:09:43.283496418 +0000 UTC m=+883.184088123" watchObservedRunningTime="2025-11-24 07:09:43.283528158 +0000 UTC m=+883.184119863" Nov 24 07:09:47 crc kubenswrapper[4809]: I1124 07:09:47.293130 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tnwx9" Nov 24 07:09:47 crc kubenswrapper[4809]: I1124 07:09:47.345003 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tnwx9" Nov 24 07:09:47 crc kubenswrapper[4809]: I1124 07:09:47.535668 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tnwx9"] Nov 24 07:09:48 crc kubenswrapper[4809]: I1124 07:09:48.042933 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:09:48 crc kubenswrapper[4809]: I1124 07:09:48.043039 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:09:49 crc kubenswrapper[4809]: I1124 07:09:49.304194 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tnwx9" podUID="da3c2694-88ce-4440-a0de-f63475eff3cf" containerName="registry-server" containerID="cri-o://9e5167351e794b8ebfad543a0b0ff334e89bfde5e74d8feb40e8f51a5bbf6217" gracePeriod=2 Nov 24 07:09:50 crc kubenswrapper[4809]: I1124 07:09:50.314138 4809 generic.go:334] "Generic (PLEG): container finished" podID="da3c2694-88ce-4440-a0de-f63475eff3cf" containerID="9e5167351e794b8ebfad543a0b0ff334e89bfde5e74d8feb40e8f51a5bbf6217" exitCode=0 Nov 24 07:09:50 crc kubenswrapper[4809]: I1124 07:09:50.314359 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tnwx9" event={"ID":"da3c2694-88ce-4440-a0de-f63475eff3cf","Type":"ContainerDied","Data":"9e5167351e794b8ebfad543a0b0ff334e89bfde5e74d8feb40e8f51a5bbf6217"} Nov 24 07:09:50 crc kubenswrapper[4809]: I1124 07:09:50.610537 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-twjz4" Nov 24 07:09:52 crc kubenswrapper[4809]: I1124 07:09:52.293415 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tnwx9" Nov 24 07:09:52 crc kubenswrapper[4809]: I1124 07:09:52.334552 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tnwx9" event={"ID":"da3c2694-88ce-4440-a0de-f63475eff3cf","Type":"ContainerDied","Data":"1214be828e40a40b5c1fca3e9108314953c2c1cebe843c49c5c1afd694656ff3"} Nov 24 07:09:52 crc kubenswrapper[4809]: I1124 07:09:52.334602 4809 scope.go:117] "RemoveContainer" containerID="9e5167351e794b8ebfad543a0b0ff334e89bfde5e74d8feb40e8f51a5bbf6217" Nov 24 07:09:52 crc kubenswrapper[4809]: I1124 07:09:52.334646 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tnwx9" Nov 24 07:09:52 crc kubenswrapper[4809]: I1124 07:09:52.347873 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da3c2694-88ce-4440-a0de-f63475eff3cf-utilities\") pod \"da3c2694-88ce-4440-a0de-f63475eff3cf\" (UID: \"da3c2694-88ce-4440-a0de-f63475eff3cf\") " Nov 24 07:09:52 crc kubenswrapper[4809]: I1124 07:09:52.348126 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpj9r\" (UniqueName: \"kubernetes.io/projected/da3c2694-88ce-4440-a0de-f63475eff3cf-kube-api-access-hpj9r\") pod \"da3c2694-88ce-4440-a0de-f63475eff3cf\" (UID: \"da3c2694-88ce-4440-a0de-f63475eff3cf\") " Nov 24 07:09:52 crc kubenswrapper[4809]: I1124 07:09:52.348259 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da3c2694-88ce-4440-a0de-f63475eff3cf-catalog-content\") pod \"da3c2694-88ce-4440-a0de-f63475eff3cf\" (UID: \"da3c2694-88ce-4440-a0de-f63475eff3cf\") " Nov 24 07:09:52 crc kubenswrapper[4809]: I1124 07:09:52.349422 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da3c2694-88ce-4440-a0de-f63475eff3cf-utilities" (OuterVolumeSpecName: "utilities") pod "da3c2694-88ce-4440-a0de-f63475eff3cf" (UID: "da3c2694-88ce-4440-a0de-f63475eff3cf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:52 crc kubenswrapper[4809]: I1124 07:09:52.357630 4809 scope.go:117] "RemoveContainer" containerID="bc284e794dd87a9d1d7c1f50c5ab0bff28956537b6ac2de3983e9078149e7d0a" Nov 24 07:09:52 crc kubenswrapper[4809]: I1124 07:09:52.365314 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da3c2694-88ce-4440-a0de-f63475eff3cf-kube-api-access-hpj9r" (OuterVolumeSpecName: "kube-api-access-hpj9r") pod "da3c2694-88ce-4440-a0de-f63475eff3cf" (UID: "da3c2694-88ce-4440-a0de-f63475eff3cf"). InnerVolumeSpecName "kube-api-access-hpj9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:52 crc kubenswrapper[4809]: I1124 07:09:52.403221 4809 scope.go:117] "RemoveContainer" containerID="c0f5bd32a98731969fcb35d06511594d307641369b6619ce0b13199ba97b6362" Nov 24 07:09:52 crc kubenswrapper[4809]: I1124 07:09:52.450563 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da3c2694-88ce-4440-a0de-f63475eff3cf-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:52 crc kubenswrapper[4809]: I1124 07:09:52.450610 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpj9r\" (UniqueName: \"kubernetes.io/projected/da3c2694-88ce-4440-a0de-f63475eff3cf-kube-api-access-hpj9r\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:52 crc kubenswrapper[4809]: I1124 07:09:52.452494 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da3c2694-88ce-4440-a0de-f63475eff3cf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "da3c2694-88ce-4440-a0de-f63475eff3cf" (UID: "da3c2694-88ce-4440-a0de-f63475eff3cf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:52 crc kubenswrapper[4809]: I1124 07:09:52.551764 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da3c2694-88ce-4440-a0de-f63475eff3cf-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:52 crc kubenswrapper[4809]: I1124 07:09:52.685427 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tnwx9"] Nov 24 07:09:52 crc kubenswrapper[4809]: I1124 07:09:52.689690 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tnwx9"] Nov 24 07:09:52 crc kubenswrapper[4809]: I1124 07:09:52.901280 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da3c2694-88ce-4440-a0de-f63475eff3cf" path="/var/lib/kubelet/pods/da3c2694-88ce-4440-a0de-f63475eff3cf/volumes" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.454371 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-csk58"] Nov 24 07:10:07 crc kubenswrapper[4809]: E1124 07:10:07.456433 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da3c2694-88ce-4440-a0de-f63475eff3cf" containerName="registry-server" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.456711 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="da3c2694-88ce-4440-a0de-f63475eff3cf" containerName="registry-server" Nov 24 07:10:07 crc kubenswrapper[4809]: E1124 07:10:07.456827 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da3c2694-88ce-4440-a0de-f63475eff3cf" containerName="extract-utilities" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.456911 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="da3c2694-88ce-4440-a0de-f63475eff3cf" containerName="extract-utilities" Nov 24 07:10:07 crc kubenswrapper[4809]: E1124 07:10:07.457017 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da3c2694-88ce-4440-a0de-f63475eff3cf" containerName="extract-content" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.457108 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="da3c2694-88ce-4440-a0de-f63475eff3cf" containerName="extract-content" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.457375 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="da3c2694-88ce-4440-a0de-f63475eff3cf" containerName="registry-server" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.458610 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-csk58" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.460493 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.460526 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.461077 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-dfjpz" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.461354 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.465908 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-csk58"] Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.521008 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-dbv92"] Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.522229 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-dbv92" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.525499 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.537298 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-dbv92"] Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.658561 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8430d1e7-0ad2-4e98-aec8-e4fb4da197fe-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-dbv92\" (UID: \"8430d1e7-0ad2-4e98-aec8-e4fb4da197fe\") " pod="openstack/dnsmasq-dns-78dd6ddcc-dbv92" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.658613 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtfdq\" (UniqueName: \"kubernetes.io/projected/8430d1e7-0ad2-4e98-aec8-e4fb4da197fe-kube-api-access-dtfdq\") pod \"dnsmasq-dns-78dd6ddcc-dbv92\" (UID: \"8430d1e7-0ad2-4e98-aec8-e4fb4da197fe\") " pod="openstack/dnsmasq-dns-78dd6ddcc-dbv92" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.658656 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90346ebb-804f-449b-a13b-dab0fe49ffa2-config\") pod \"dnsmasq-dns-675f4bcbfc-csk58\" (UID: \"90346ebb-804f-449b-a13b-dab0fe49ffa2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-csk58" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.658685 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcnlk\" (UniqueName: \"kubernetes.io/projected/90346ebb-804f-449b-a13b-dab0fe49ffa2-kube-api-access-wcnlk\") pod \"dnsmasq-dns-675f4bcbfc-csk58\" (UID: \"90346ebb-804f-449b-a13b-dab0fe49ffa2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-csk58" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.658714 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8430d1e7-0ad2-4e98-aec8-e4fb4da197fe-config\") pod \"dnsmasq-dns-78dd6ddcc-dbv92\" (UID: \"8430d1e7-0ad2-4e98-aec8-e4fb4da197fe\") " pod="openstack/dnsmasq-dns-78dd6ddcc-dbv92" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.759711 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8430d1e7-0ad2-4e98-aec8-e4fb4da197fe-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-dbv92\" (UID: \"8430d1e7-0ad2-4e98-aec8-e4fb4da197fe\") " pod="openstack/dnsmasq-dns-78dd6ddcc-dbv92" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.759758 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtfdq\" (UniqueName: \"kubernetes.io/projected/8430d1e7-0ad2-4e98-aec8-e4fb4da197fe-kube-api-access-dtfdq\") pod \"dnsmasq-dns-78dd6ddcc-dbv92\" (UID: \"8430d1e7-0ad2-4e98-aec8-e4fb4da197fe\") " pod="openstack/dnsmasq-dns-78dd6ddcc-dbv92" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.759797 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90346ebb-804f-449b-a13b-dab0fe49ffa2-config\") pod \"dnsmasq-dns-675f4bcbfc-csk58\" (UID: \"90346ebb-804f-449b-a13b-dab0fe49ffa2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-csk58" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.759825 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcnlk\" (UniqueName: \"kubernetes.io/projected/90346ebb-804f-449b-a13b-dab0fe49ffa2-kube-api-access-wcnlk\") pod \"dnsmasq-dns-675f4bcbfc-csk58\" (UID: \"90346ebb-804f-449b-a13b-dab0fe49ffa2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-csk58" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.759854 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8430d1e7-0ad2-4e98-aec8-e4fb4da197fe-config\") pod \"dnsmasq-dns-78dd6ddcc-dbv92\" (UID: \"8430d1e7-0ad2-4e98-aec8-e4fb4da197fe\") " pod="openstack/dnsmasq-dns-78dd6ddcc-dbv92" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.760633 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8430d1e7-0ad2-4e98-aec8-e4fb4da197fe-config\") pod \"dnsmasq-dns-78dd6ddcc-dbv92\" (UID: \"8430d1e7-0ad2-4e98-aec8-e4fb4da197fe\") " pod="openstack/dnsmasq-dns-78dd6ddcc-dbv92" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.761122 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8430d1e7-0ad2-4e98-aec8-e4fb4da197fe-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-dbv92\" (UID: \"8430d1e7-0ad2-4e98-aec8-e4fb4da197fe\") " pod="openstack/dnsmasq-dns-78dd6ddcc-dbv92" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.762321 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90346ebb-804f-449b-a13b-dab0fe49ffa2-config\") pod \"dnsmasq-dns-675f4bcbfc-csk58\" (UID: \"90346ebb-804f-449b-a13b-dab0fe49ffa2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-csk58" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.779274 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtfdq\" (UniqueName: \"kubernetes.io/projected/8430d1e7-0ad2-4e98-aec8-e4fb4da197fe-kube-api-access-dtfdq\") pod \"dnsmasq-dns-78dd6ddcc-dbv92\" (UID: \"8430d1e7-0ad2-4e98-aec8-e4fb4da197fe\") " pod="openstack/dnsmasq-dns-78dd6ddcc-dbv92" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.779708 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcnlk\" (UniqueName: \"kubernetes.io/projected/90346ebb-804f-449b-a13b-dab0fe49ffa2-kube-api-access-wcnlk\") pod \"dnsmasq-dns-675f4bcbfc-csk58\" (UID: \"90346ebb-804f-449b-a13b-dab0fe49ffa2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-csk58" Nov 24 07:10:07 crc kubenswrapper[4809]: I1124 07:10:07.838850 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-dbv92" Nov 24 07:10:08 crc kubenswrapper[4809]: I1124 07:10:08.079318 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-csk58" Nov 24 07:10:08 crc kubenswrapper[4809]: I1124 07:10:08.259614 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-dbv92"] Nov 24 07:10:08 crc kubenswrapper[4809]: I1124 07:10:08.262513 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 07:10:08 crc kubenswrapper[4809]: I1124 07:10:08.477204 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-csk58"] Nov 24 07:10:08 crc kubenswrapper[4809]: W1124 07:10:08.481526 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90346ebb_804f_449b_a13b_dab0fe49ffa2.slice/crio-74f1cab12fccf0ce837d8f52b995befab27b485313681a938049d69100aa266a WatchSource:0}: Error finding container 74f1cab12fccf0ce837d8f52b995befab27b485313681a938049d69100aa266a: Status 404 returned error can't find the container with id 74f1cab12fccf0ce837d8f52b995befab27b485313681a938049d69100aa266a Nov 24 07:10:08 crc kubenswrapper[4809]: I1124 07:10:08.491327 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-dbv92" event={"ID":"8430d1e7-0ad2-4e98-aec8-e4fb4da197fe","Type":"ContainerStarted","Data":"cfaaf47c03c98bd962e63d1cf23909282b1c619bada839803bbea8ca98bcd1f4"} Nov 24 07:10:08 crc kubenswrapper[4809]: I1124 07:10:08.493009 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-csk58" event={"ID":"90346ebb-804f-449b-a13b-dab0fe49ffa2","Type":"ContainerStarted","Data":"74f1cab12fccf0ce837d8f52b995befab27b485313681a938049d69100aa266a"} Nov 24 07:10:10 crc kubenswrapper[4809]: I1124 07:10:10.697859 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-csk58"] Nov 24 07:10:10 crc kubenswrapper[4809]: I1124 07:10:10.716019 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jslst"] Nov 24 07:10:10 crc kubenswrapper[4809]: I1124 07:10:10.718872 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jslst" Nov 24 07:10:10 crc kubenswrapper[4809]: I1124 07:10:10.741935 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jslst"] Nov 24 07:10:10 crc kubenswrapper[4809]: I1124 07:10:10.926768 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1041a99c-ae53-4ec0-9f44-bc33025d8566-config\") pod \"dnsmasq-dns-666b6646f7-jslst\" (UID: \"1041a99c-ae53-4ec0-9f44-bc33025d8566\") " pod="openstack/dnsmasq-dns-666b6646f7-jslst" Nov 24 07:10:10 crc kubenswrapper[4809]: I1124 07:10:10.926852 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1041a99c-ae53-4ec0-9f44-bc33025d8566-dns-svc\") pod \"dnsmasq-dns-666b6646f7-jslst\" (UID: \"1041a99c-ae53-4ec0-9f44-bc33025d8566\") " pod="openstack/dnsmasq-dns-666b6646f7-jslst" Nov 24 07:10:10 crc kubenswrapper[4809]: I1124 07:10:10.926892 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mgff\" (UniqueName: \"kubernetes.io/projected/1041a99c-ae53-4ec0-9f44-bc33025d8566-kube-api-access-9mgff\") pod \"dnsmasq-dns-666b6646f7-jslst\" (UID: \"1041a99c-ae53-4ec0-9f44-bc33025d8566\") " pod="openstack/dnsmasq-dns-666b6646f7-jslst" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.000278 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-dbv92"] Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.027427 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dzjmv"] Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.028907 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-dzjmv" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.031892 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1041a99c-ae53-4ec0-9f44-bc33025d8566-dns-svc\") pod \"dnsmasq-dns-666b6646f7-jslst\" (UID: \"1041a99c-ae53-4ec0-9f44-bc33025d8566\") " pod="openstack/dnsmasq-dns-666b6646f7-jslst" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.031948 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbcae289-c075-4ecf-997d-64f15f6a8fbf-config\") pod \"dnsmasq-dns-57d769cc4f-dzjmv\" (UID: \"dbcae289-c075-4ecf-997d-64f15f6a8fbf\") " pod="openstack/dnsmasq-dns-57d769cc4f-dzjmv" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.031986 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nbkk\" (UniqueName: \"kubernetes.io/projected/dbcae289-c075-4ecf-997d-64f15f6a8fbf-kube-api-access-6nbkk\") pod \"dnsmasq-dns-57d769cc4f-dzjmv\" (UID: \"dbcae289-c075-4ecf-997d-64f15f6a8fbf\") " pod="openstack/dnsmasq-dns-57d769cc4f-dzjmv" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.032010 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mgff\" (UniqueName: \"kubernetes.io/projected/1041a99c-ae53-4ec0-9f44-bc33025d8566-kube-api-access-9mgff\") pod \"dnsmasq-dns-666b6646f7-jslst\" (UID: \"1041a99c-ae53-4ec0-9f44-bc33025d8566\") " pod="openstack/dnsmasq-dns-666b6646f7-jslst" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.032039 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1041a99c-ae53-4ec0-9f44-bc33025d8566-config\") pod \"dnsmasq-dns-666b6646f7-jslst\" (UID: \"1041a99c-ae53-4ec0-9f44-bc33025d8566\") " pod="openstack/dnsmasq-dns-666b6646f7-jslst" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.032109 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dbcae289-c075-4ecf-997d-64f15f6a8fbf-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-dzjmv\" (UID: \"dbcae289-c075-4ecf-997d-64f15f6a8fbf\") " pod="openstack/dnsmasq-dns-57d769cc4f-dzjmv" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.033922 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1041a99c-ae53-4ec0-9f44-bc33025d8566-config\") pod \"dnsmasq-dns-666b6646f7-jslst\" (UID: \"1041a99c-ae53-4ec0-9f44-bc33025d8566\") " pod="openstack/dnsmasq-dns-666b6646f7-jslst" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.033936 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1041a99c-ae53-4ec0-9f44-bc33025d8566-dns-svc\") pod \"dnsmasq-dns-666b6646f7-jslst\" (UID: \"1041a99c-ae53-4ec0-9f44-bc33025d8566\") " pod="openstack/dnsmasq-dns-666b6646f7-jslst" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.044114 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dzjmv"] Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.058593 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mgff\" (UniqueName: \"kubernetes.io/projected/1041a99c-ae53-4ec0-9f44-bc33025d8566-kube-api-access-9mgff\") pod \"dnsmasq-dns-666b6646f7-jslst\" (UID: \"1041a99c-ae53-4ec0-9f44-bc33025d8566\") " pod="openstack/dnsmasq-dns-666b6646f7-jslst" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.132797 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dbcae289-c075-4ecf-997d-64f15f6a8fbf-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-dzjmv\" (UID: \"dbcae289-c075-4ecf-997d-64f15f6a8fbf\") " pod="openstack/dnsmasq-dns-57d769cc4f-dzjmv" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.132860 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbcae289-c075-4ecf-997d-64f15f6a8fbf-config\") pod \"dnsmasq-dns-57d769cc4f-dzjmv\" (UID: \"dbcae289-c075-4ecf-997d-64f15f6a8fbf\") " pod="openstack/dnsmasq-dns-57d769cc4f-dzjmv" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.132884 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nbkk\" (UniqueName: \"kubernetes.io/projected/dbcae289-c075-4ecf-997d-64f15f6a8fbf-kube-api-access-6nbkk\") pod \"dnsmasq-dns-57d769cc4f-dzjmv\" (UID: \"dbcae289-c075-4ecf-997d-64f15f6a8fbf\") " pod="openstack/dnsmasq-dns-57d769cc4f-dzjmv" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.133660 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dbcae289-c075-4ecf-997d-64f15f6a8fbf-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-dzjmv\" (UID: \"dbcae289-c075-4ecf-997d-64f15f6a8fbf\") " pod="openstack/dnsmasq-dns-57d769cc4f-dzjmv" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.133935 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbcae289-c075-4ecf-997d-64f15f6a8fbf-config\") pod \"dnsmasq-dns-57d769cc4f-dzjmv\" (UID: \"dbcae289-c075-4ecf-997d-64f15f6a8fbf\") " pod="openstack/dnsmasq-dns-57d769cc4f-dzjmv" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.151533 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nbkk\" (UniqueName: \"kubernetes.io/projected/dbcae289-c075-4ecf-997d-64f15f6a8fbf-kube-api-access-6nbkk\") pod \"dnsmasq-dns-57d769cc4f-dzjmv\" (UID: \"dbcae289-c075-4ecf-997d-64f15f6a8fbf\") " pod="openstack/dnsmasq-dns-57d769cc4f-dzjmv" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.338109 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jslst" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.354211 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-dzjmv" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.843255 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.844317 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.846821 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.847019 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.847179 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.847320 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.848604 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.848747 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.854452 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:10:11 crc kubenswrapper[4809]: I1124 07:10:11.854942 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-xfhq2" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.045738 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.045787 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7d0a537c-b1f0-4908-a604-424ea6a58438-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.045822 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.046043 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.046171 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7d0a537c-b1f0-4908-a604-424ea6a58438-config-data\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.046194 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7d0a537c-b1f0-4908-a604-424ea6a58438-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.046223 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7d0a537c-b1f0-4908-a604-424ea6a58438-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.046245 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7cjz\" (UniqueName: \"kubernetes.io/projected/7d0a537c-b1f0-4908-a604-424ea6a58438-kube-api-access-w7cjz\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.046344 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.046425 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7d0a537c-b1f0-4908-a604-424ea6a58438-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.046458 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.148445 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.148513 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.148570 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7d0a537c-b1f0-4908-a604-424ea6a58438-config-data\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.148589 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7d0a537c-b1f0-4908-a604-424ea6a58438-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.148616 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7d0a537c-b1f0-4908-a604-424ea6a58438-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.148638 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7cjz\" (UniqueName: \"kubernetes.io/projected/7d0a537c-b1f0-4908-a604-424ea6a58438-kube-api-access-w7cjz\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.148676 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.148708 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7d0a537c-b1f0-4908-a604-424ea6a58438-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.148731 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.148749 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.148764 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7d0a537c-b1f0-4908-a604-424ea6a58438-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.148865 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.149393 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.149930 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7d0a537c-b1f0-4908-a604-424ea6a58438-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.150227 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7d0a537c-b1f0-4908-a604-424ea6a58438-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.150233 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7d0a537c-b1f0-4908-a604-424ea6a58438-config-data\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.154048 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7d0a537c-b1f0-4908-a604-424ea6a58438-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.154136 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.157640 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7d0a537c-b1f0-4908-a604-424ea6a58438-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.158455 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.162879 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.172593 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7cjz\" (UniqueName: \"kubernetes.io/projected/7d0a537c-b1f0-4908-a604-424ea6a58438-kube-api-access-w7cjz\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.174809 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.176655 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.185624 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.189935 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.191443 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.191554 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.191646 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.192254 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.192504 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-ncq9f" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.192529 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.192699 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.352349 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.352463 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnlc6\" (UniqueName: \"kubernetes.io/projected/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-kube-api-access-lnlc6\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.352508 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.352548 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.352589 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.352646 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.352690 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.352743 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.352779 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.352832 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.352863 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.453594 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.453705 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.453726 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnlc6\" (UniqueName: \"kubernetes.io/projected/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-kube-api-access-lnlc6\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.453743 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.453764 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.453785 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.454576 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.454595 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.454654 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.454886 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.454933 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.454705 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.457124 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.457152 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.457187 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.457699 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.457986 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.461182 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.462643 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.466073 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.466214 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.467021 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.479100 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnlc6\" (UniqueName: \"kubernetes.io/projected/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-kube-api-access-lnlc6\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.502211 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:12 crc kubenswrapper[4809]: I1124 07:10:12.551725 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.559517 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.561413 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.563510 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.563591 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-nlsdv" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.563762 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.565878 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.569349 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.569859 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.674011 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.674054 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caea018d-04b5-4d20-8161-9a9a87f67583-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.674097 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/caea018d-04b5-4d20-8161-9a9a87f67583-config-data-generated\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.674126 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/caea018d-04b5-4d20-8161-9a9a87f67583-operator-scripts\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.674253 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/caea018d-04b5-4d20-8161-9a9a87f67583-kolla-config\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.674291 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/caea018d-04b5-4d20-8161-9a9a87f67583-config-data-default\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.674314 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97xk2\" (UniqueName: \"kubernetes.io/projected/caea018d-04b5-4d20-8161-9a9a87f67583-kube-api-access-97xk2\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.674354 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/caea018d-04b5-4d20-8161-9a9a87f67583-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.776172 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/caea018d-04b5-4d20-8161-9a9a87f67583-kolla-config\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.776220 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/caea018d-04b5-4d20-8161-9a9a87f67583-config-data-default\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.776267 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97xk2\" (UniqueName: \"kubernetes.io/projected/caea018d-04b5-4d20-8161-9a9a87f67583-kube-api-access-97xk2\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.776309 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/caea018d-04b5-4d20-8161-9a9a87f67583-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.776724 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.776749 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caea018d-04b5-4d20-8161-9a9a87f67583-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.776930 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/caea018d-04b5-4d20-8161-9a9a87f67583-kolla-config\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.776980 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.777144 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/caea018d-04b5-4d20-8161-9a9a87f67583-config-data-generated\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.777235 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/caea018d-04b5-4d20-8161-9a9a87f67583-operator-scripts\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.777298 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/caea018d-04b5-4d20-8161-9a9a87f67583-config-data-default\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.777471 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/caea018d-04b5-4d20-8161-9a9a87f67583-config-data-generated\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.778466 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/caea018d-04b5-4d20-8161-9a9a87f67583-operator-scripts\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.782147 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caea018d-04b5-4d20-8161-9a9a87f67583-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.782280 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/caea018d-04b5-4d20-8161-9a9a87f67583-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.794324 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97xk2\" (UniqueName: \"kubernetes.io/projected/caea018d-04b5-4d20-8161-9a9a87f67583-kube-api-access-97xk2\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.800914 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"caea018d-04b5-4d20-8161-9a9a87f67583\") " pod="openstack/openstack-galera-0" Nov 24 07:10:13 crc kubenswrapper[4809]: I1124 07:10:13.888904 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 07:10:14 crc kubenswrapper[4809]: I1124 07:10:14.956353 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 07:10:14 crc kubenswrapper[4809]: I1124 07:10:14.957620 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:14 crc kubenswrapper[4809]: I1124 07:10:14.959680 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 24 07:10:14 crc kubenswrapper[4809]: I1124 07:10:14.960019 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-9w7lf" Nov 24 07:10:14 crc kubenswrapper[4809]: I1124 07:10:14.960264 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 24 07:10:14 crc kubenswrapper[4809]: I1124 07:10:14.960397 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 24 07:10:14 crc kubenswrapper[4809]: I1124 07:10:14.975275 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.097260 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.097520 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5afba7d-dbf3-4aa6-863b-1d3a6d71af88-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.097575 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e5afba7d-dbf3-4aa6-863b-1d3a6d71af88-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.097617 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5afba7d-dbf3-4aa6-863b-1d3a6d71af88-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.097791 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e5afba7d-dbf3-4aa6-863b-1d3a6d71af88-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.097854 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8xq7\" (UniqueName: \"kubernetes.io/projected/e5afba7d-dbf3-4aa6-863b-1d3a6d71af88-kube-api-access-t8xq7\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.097893 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5afba7d-dbf3-4aa6-863b-1d3a6d71af88-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.097936 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e5afba7d-dbf3-4aa6-863b-1d3a6d71af88-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.198877 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e5afba7d-dbf3-4aa6-863b-1d3a6d71af88-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.198941 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8xq7\" (UniqueName: \"kubernetes.io/projected/e5afba7d-dbf3-4aa6-863b-1d3a6d71af88-kube-api-access-t8xq7\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.198984 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5afba7d-dbf3-4aa6-863b-1d3a6d71af88-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.199016 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e5afba7d-dbf3-4aa6-863b-1d3a6d71af88-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.199073 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.199095 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5afba7d-dbf3-4aa6-863b-1d3a6d71af88-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.199153 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e5afba7d-dbf3-4aa6-863b-1d3a6d71af88-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.199202 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5afba7d-dbf3-4aa6-863b-1d3a6d71af88-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.200436 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e5afba7d-dbf3-4aa6-863b-1d3a6d71af88-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.200449 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.201427 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e5afba7d-dbf3-4aa6-863b-1d3a6d71af88-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.202454 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e5afba7d-dbf3-4aa6-863b-1d3a6d71af88-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.204397 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5afba7d-dbf3-4aa6-863b-1d3a6d71af88-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.205086 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5afba7d-dbf3-4aa6-863b-1d3a6d71af88-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.217515 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5afba7d-dbf3-4aa6-863b-1d3a6d71af88-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.229082 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.241537 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8xq7\" (UniqueName: \"kubernetes.io/projected/e5afba7d-dbf3-4aa6-863b-1d3a6d71af88-kube-api-access-t8xq7\") pod \"openstack-cell1-galera-0\" (UID: \"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.284550 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.393720 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.394655 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.398019 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.400491 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.400695 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-4z947" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.420323 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.502798 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/ced578a4-f9b1-4d1f-978f-beda4b33bc59-memcached-tls-certs\") pod \"memcached-0\" (UID: \"ced578a4-f9b1-4d1f-978f-beda4b33bc59\") " pod="openstack/memcached-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.502895 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ced578a4-f9b1-4d1f-978f-beda4b33bc59-config-data\") pod \"memcached-0\" (UID: \"ced578a4-f9b1-4d1f-978f-beda4b33bc59\") " pod="openstack/memcached-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.502941 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ced578a4-f9b1-4d1f-978f-beda4b33bc59-kolla-config\") pod \"memcached-0\" (UID: \"ced578a4-f9b1-4d1f-978f-beda4b33bc59\") " pod="openstack/memcached-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.502958 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ced578a4-f9b1-4d1f-978f-beda4b33bc59-combined-ca-bundle\") pod \"memcached-0\" (UID: \"ced578a4-f9b1-4d1f-978f-beda4b33bc59\") " pod="openstack/memcached-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.502991 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9dkq\" (UniqueName: \"kubernetes.io/projected/ced578a4-f9b1-4d1f-978f-beda4b33bc59-kube-api-access-p9dkq\") pod \"memcached-0\" (UID: \"ced578a4-f9b1-4d1f-978f-beda4b33bc59\") " pod="openstack/memcached-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.605017 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/ced578a4-f9b1-4d1f-978f-beda4b33bc59-memcached-tls-certs\") pod \"memcached-0\" (UID: \"ced578a4-f9b1-4d1f-978f-beda4b33bc59\") " pod="openstack/memcached-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.605073 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ced578a4-f9b1-4d1f-978f-beda4b33bc59-config-data\") pod \"memcached-0\" (UID: \"ced578a4-f9b1-4d1f-978f-beda4b33bc59\") " pod="openstack/memcached-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.605128 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ced578a4-f9b1-4d1f-978f-beda4b33bc59-kolla-config\") pod \"memcached-0\" (UID: \"ced578a4-f9b1-4d1f-978f-beda4b33bc59\") " pod="openstack/memcached-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.605146 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ced578a4-f9b1-4d1f-978f-beda4b33bc59-combined-ca-bundle\") pod \"memcached-0\" (UID: \"ced578a4-f9b1-4d1f-978f-beda4b33bc59\") " pod="openstack/memcached-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.605167 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9dkq\" (UniqueName: \"kubernetes.io/projected/ced578a4-f9b1-4d1f-978f-beda4b33bc59-kube-api-access-p9dkq\") pod \"memcached-0\" (UID: \"ced578a4-f9b1-4d1f-978f-beda4b33bc59\") " pod="openstack/memcached-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.606828 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ced578a4-f9b1-4d1f-978f-beda4b33bc59-config-data\") pod \"memcached-0\" (UID: \"ced578a4-f9b1-4d1f-978f-beda4b33bc59\") " pod="openstack/memcached-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.607602 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ced578a4-f9b1-4d1f-978f-beda4b33bc59-kolla-config\") pod \"memcached-0\" (UID: \"ced578a4-f9b1-4d1f-978f-beda4b33bc59\") " pod="openstack/memcached-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.609238 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/ced578a4-f9b1-4d1f-978f-beda4b33bc59-memcached-tls-certs\") pod \"memcached-0\" (UID: \"ced578a4-f9b1-4d1f-978f-beda4b33bc59\") " pod="openstack/memcached-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.618677 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ced578a4-f9b1-4d1f-978f-beda4b33bc59-combined-ca-bundle\") pod \"memcached-0\" (UID: \"ced578a4-f9b1-4d1f-978f-beda4b33bc59\") " pod="openstack/memcached-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.633729 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9dkq\" (UniqueName: \"kubernetes.io/projected/ced578a4-f9b1-4d1f-978f-beda4b33bc59-kube-api-access-p9dkq\") pod \"memcached-0\" (UID: \"ced578a4-f9b1-4d1f-978f-beda4b33bc59\") " pod="openstack/memcached-0" Nov 24 07:10:15 crc kubenswrapper[4809]: I1124 07:10:15.710584 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 07:10:16 crc kubenswrapper[4809]: I1124 07:10:16.976580 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 07:10:16 crc kubenswrapper[4809]: I1124 07:10:16.977719 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 07:10:16 crc kubenswrapper[4809]: I1124 07:10:16.981132 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-h48ms" Nov 24 07:10:16 crc kubenswrapper[4809]: I1124 07:10:16.991398 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 07:10:17 crc kubenswrapper[4809]: I1124 07:10:17.127208 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d27cj\" (UniqueName: \"kubernetes.io/projected/652f6c1f-9038-4993-a8c2-fb3e9a6a822f-kube-api-access-d27cj\") pod \"kube-state-metrics-0\" (UID: \"652f6c1f-9038-4993-a8c2-fb3e9a6a822f\") " pod="openstack/kube-state-metrics-0" Nov 24 07:10:17 crc kubenswrapper[4809]: I1124 07:10:17.228137 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d27cj\" (UniqueName: \"kubernetes.io/projected/652f6c1f-9038-4993-a8c2-fb3e9a6a822f-kube-api-access-d27cj\") pod \"kube-state-metrics-0\" (UID: \"652f6c1f-9038-4993-a8c2-fb3e9a6a822f\") " pod="openstack/kube-state-metrics-0" Nov 24 07:10:17 crc kubenswrapper[4809]: I1124 07:10:17.267358 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d27cj\" (UniqueName: \"kubernetes.io/projected/652f6c1f-9038-4993-a8c2-fb3e9a6a822f-kube-api-access-d27cj\") pod \"kube-state-metrics-0\" (UID: \"652f6c1f-9038-4993-a8c2-fb3e9a6a822f\") " pod="openstack/kube-state-metrics-0" Nov 24 07:10:17 crc kubenswrapper[4809]: I1124 07:10:17.290372 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 07:10:18 crc kubenswrapper[4809]: I1124 07:10:18.043330 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:10:18 crc kubenswrapper[4809]: I1124 07:10:18.043456 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:10:20 crc kubenswrapper[4809]: I1124 07:10:20.861859 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-vnpwc"] Nov 24 07:10:20 crc kubenswrapper[4809]: I1124 07:10:20.863627 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:20 crc kubenswrapper[4809]: I1124 07:10:20.868217 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 24 07:10:20 crc kubenswrapper[4809]: I1124 07:10:20.869057 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-g77bx" Nov 24 07:10:20 crc kubenswrapper[4809]: I1124 07:10:20.869132 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 24 07:10:20 crc kubenswrapper[4809]: I1124 07:10:20.879776 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vnpwc"] Nov 24 07:10:20 crc kubenswrapper[4809]: I1124 07:10:20.886587 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-drz48"] Nov 24 07:10:20 crc kubenswrapper[4809]: I1124 07:10:20.888582 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:10:20 crc kubenswrapper[4809]: I1124 07:10:20.915236 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-drz48"] Nov 24 07:10:20 crc kubenswrapper[4809]: I1124 07:10:20.985807 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3c2d3d10-789d-4a2d-9926-cba915a127df-var-log\") pod \"ovn-controller-ovs-drz48\" (UID: \"3c2d3d10-789d-4a2d-9926-cba915a127df\") " pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:10:20 crc kubenswrapper[4809]: I1124 07:10:20.985856 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3c2d3d10-789d-4a2d-9926-cba915a127df-etc-ovs\") pod \"ovn-controller-ovs-drz48\" (UID: \"3c2d3d10-789d-4a2d-9926-cba915a127df\") " pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:10:20 crc kubenswrapper[4809]: I1124 07:10:20.985888 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3c2d3d10-789d-4a2d-9926-cba915a127df-var-run\") pod \"ovn-controller-ovs-drz48\" (UID: \"3c2d3d10-789d-4a2d-9926-cba915a127df\") " pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:10:20 crc kubenswrapper[4809]: I1124 07:10:20.985920 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq99x\" (UniqueName: \"kubernetes.io/projected/c14ae567-41ab-455a-8756-04daf7c81677-kube-api-access-jq99x\") pod \"ovn-controller-vnpwc\" (UID: \"c14ae567-41ab-455a-8756-04daf7c81677\") " pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:20 crc kubenswrapper[4809]: I1124 07:10:20.985942 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c14ae567-41ab-455a-8756-04daf7c81677-ovn-controller-tls-certs\") pod \"ovn-controller-vnpwc\" (UID: \"c14ae567-41ab-455a-8756-04daf7c81677\") " pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:20 crc kubenswrapper[4809]: I1124 07:10:20.985979 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c14ae567-41ab-455a-8756-04daf7c81677-var-run\") pod \"ovn-controller-vnpwc\" (UID: \"c14ae567-41ab-455a-8756-04daf7c81677\") " pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:20 crc kubenswrapper[4809]: I1124 07:10:20.986002 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c14ae567-41ab-455a-8756-04daf7c81677-var-log-ovn\") pod \"ovn-controller-vnpwc\" (UID: \"c14ae567-41ab-455a-8756-04daf7c81677\") " pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:20 crc kubenswrapper[4809]: I1124 07:10:20.986057 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c14ae567-41ab-455a-8756-04daf7c81677-var-run-ovn\") pod \"ovn-controller-vnpwc\" (UID: \"c14ae567-41ab-455a-8756-04daf7c81677\") " pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:20 crc kubenswrapper[4809]: I1124 07:10:20.986201 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khc2v\" (UniqueName: \"kubernetes.io/projected/3c2d3d10-789d-4a2d-9926-cba915a127df-kube-api-access-khc2v\") pod \"ovn-controller-ovs-drz48\" (UID: \"3c2d3d10-789d-4a2d-9926-cba915a127df\") " pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:10:20 crc kubenswrapper[4809]: I1124 07:10:20.986543 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3c2d3d10-789d-4a2d-9926-cba915a127df-var-lib\") pod \"ovn-controller-ovs-drz48\" (UID: \"3c2d3d10-789d-4a2d-9926-cba915a127df\") " pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:10:20 crc kubenswrapper[4809]: I1124 07:10:20.986567 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c14ae567-41ab-455a-8756-04daf7c81677-combined-ca-bundle\") pod \"ovn-controller-vnpwc\" (UID: \"c14ae567-41ab-455a-8756-04daf7c81677\") " pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:20 crc kubenswrapper[4809]: I1124 07:10:20.986609 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3c2d3d10-789d-4a2d-9926-cba915a127df-scripts\") pod \"ovn-controller-ovs-drz48\" (UID: \"3c2d3d10-789d-4a2d-9926-cba915a127df\") " pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:10:20 crc kubenswrapper[4809]: I1124 07:10:20.986637 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c14ae567-41ab-455a-8756-04daf7c81677-scripts\") pod \"ovn-controller-vnpwc\" (UID: \"c14ae567-41ab-455a-8756-04daf7c81677\") " pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.087661 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3c2d3d10-789d-4a2d-9926-cba915a127df-var-run\") pod \"ovn-controller-ovs-drz48\" (UID: \"3c2d3d10-789d-4a2d-9926-cba915a127df\") " pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.087725 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq99x\" (UniqueName: \"kubernetes.io/projected/c14ae567-41ab-455a-8756-04daf7c81677-kube-api-access-jq99x\") pod \"ovn-controller-vnpwc\" (UID: \"c14ae567-41ab-455a-8756-04daf7c81677\") " pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.087750 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c14ae567-41ab-455a-8756-04daf7c81677-ovn-controller-tls-certs\") pod \"ovn-controller-vnpwc\" (UID: \"c14ae567-41ab-455a-8756-04daf7c81677\") " pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.087772 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c14ae567-41ab-455a-8756-04daf7c81677-var-run\") pod \"ovn-controller-vnpwc\" (UID: \"c14ae567-41ab-455a-8756-04daf7c81677\") " pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.087792 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c14ae567-41ab-455a-8756-04daf7c81677-var-log-ovn\") pod \"ovn-controller-vnpwc\" (UID: \"c14ae567-41ab-455a-8756-04daf7c81677\") " pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.087823 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c14ae567-41ab-455a-8756-04daf7c81677-var-run-ovn\") pod \"ovn-controller-vnpwc\" (UID: \"c14ae567-41ab-455a-8756-04daf7c81677\") " pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.087843 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khc2v\" (UniqueName: \"kubernetes.io/projected/3c2d3d10-789d-4a2d-9926-cba915a127df-kube-api-access-khc2v\") pod \"ovn-controller-ovs-drz48\" (UID: \"3c2d3d10-789d-4a2d-9926-cba915a127df\") " pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.087864 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3c2d3d10-789d-4a2d-9926-cba915a127df-var-lib\") pod \"ovn-controller-ovs-drz48\" (UID: \"3c2d3d10-789d-4a2d-9926-cba915a127df\") " pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.087890 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c14ae567-41ab-455a-8756-04daf7c81677-combined-ca-bundle\") pod \"ovn-controller-vnpwc\" (UID: \"c14ae567-41ab-455a-8756-04daf7c81677\") " pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.087910 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3c2d3d10-789d-4a2d-9926-cba915a127df-scripts\") pod \"ovn-controller-ovs-drz48\" (UID: \"3c2d3d10-789d-4a2d-9926-cba915a127df\") " pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.087930 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c14ae567-41ab-455a-8756-04daf7c81677-scripts\") pod \"ovn-controller-vnpwc\" (UID: \"c14ae567-41ab-455a-8756-04daf7c81677\") " pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.087984 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3c2d3d10-789d-4a2d-9926-cba915a127df-var-log\") pod \"ovn-controller-ovs-drz48\" (UID: \"3c2d3d10-789d-4a2d-9926-cba915a127df\") " pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.088005 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3c2d3d10-789d-4a2d-9926-cba915a127df-etc-ovs\") pod \"ovn-controller-ovs-drz48\" (UID: \"3c2d3d10-789d-4a2d-9926-cba915a127df\") " pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.088641 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c14ae567-41ab-455a-8756-04daf7c81677-var-log-ovn\") pod \"ovn-controller-vnpwc\" (UID: \"c14ae567-41ab-455a-8756-04daf7c81677\") " pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.088719 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3c2d3d10-789d-4a2d-9926-cba915a127df-var-log\") pod \"ovn-controller-ovs-drz48\" (UID: \"3c2d3d10-789d-4a2d-9926-cba915a127df\") " pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.088763 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3c2d3d10-789d-4a2d-9926-cba915a127df-etc-ovs\") pod \"ovn-controller-ovs-drz48\" (UID: \"3c2d3d10-789d-4a2d-9926-cba915a127df\") " pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.088875 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c14ae567-41ab-455a-8756-04daf7c81677-var-run\") pod \"ovn-controller-vnpwc\" (UID: \"c14ae567-41ab-455a-8756-04daf7c81677\") " pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.089086 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3c2d3d10-789d-4a2d-9926-cba915a127df-var-run\") pod \"ovn-controller-ovs-drz48\" (UID: \"3c2d3d10-789d-4a2d-9926-cba915a127df\") " pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.089110 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3c2d3d10-789d-4a2d-9926-cba915a127df-var-lib\") pod \"ovn-controller-ovs-drz48\" (UID: \"3c2d3d10-789d-4a2d-9926-cba915a127df\") " pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.090392 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c14ae567-41ab-455a-8756-04daf7c81677-scripts\") pod \"ovn-controller-vnpwc\" (UID: \"c14ae567-41ab-455a-8756-04daf7c81677\") " pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.090529 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c14ae567-41ab-455a-8756-04daf7c81677-var-run-ovn\") pod \"ovn-controller-vnpwc\" (UID: \"c14ae567-41ab-455a-8756-04daf7c81677\") " pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.090819 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3c2d3d10-789d-4a2d-9926-cba915a127df-scripts\") pod \"ovn-controller-ovs-drz48\" (UID: \"3c2d3d10-789d-4a2d-9926-cba915a127df\") " pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.095071 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c14ae567-41ab-455a-8756-04daf7c81677-combined-ca-bundle\") pod \"ovn-controller-vnpwc\" (UID: \"c14ae567-41ab-455a-8756-04daf7c81677\") " pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.107370 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c14ae567-41ab-455a-8756-04daf7c81677-ovn-controller-tls-certs\") pod \"ovn-controller-vnpwc\" (UID: \"c14ae567-41ab-455a-8756-04daf7c81677\") " pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.112585 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq99x\" (UniqueName: \"kubernetes.io/projected/c14ae567-41ab-455a-8756-04daf7c81677-kube-api-access-jq99x\") pod \"ovn-controller-vnpwc\" (UID: \"c14ae567-41ab-455a-8756-04daf7c81677\") " pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.112699 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khc2v\" (UniqueName: \"kubernetes.io/projected/3c2d3d10-789d-4a2d-9926-cba915a127df-kube-api-access-khc2v\") pod \"ovn-controller-ovs-drz48\" (UID: \"3c2d3d10-789d-4a2d-9926-cba915a127df\") " pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.181242 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.213197 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:10:21 crc kubenswrapper[4809]: I1124 07:10:21.497325 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:10:21 crc kubenswrapper[4809]: W1124 07:10:21.927537 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf91f6ab4_0c7c_4002_aad9_dcd408c9000e.slice/crio-d661a253ac318aced47d491b5aefdb4281e8dec6c3f04fff943e9ad0ef6cbf97 WatchSource:0}: Error finding container d661a253ac318aced47d491b5aefdb4281e8dec6c3f04fff943e9ad0ef6cbf97: Status 404 returned error can't find the container with id d661a253ac318aced47d491b5aefdb4281e8dec6c3f04fff943e9ad0ef6cbf97 Nov 24 07:10:21 crc kubenswrapper[4809]: E1124 07:10:21.960039 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 07:10:21 crc kubenswrapper[4809]: E1124 07:10:21.960212 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wcnlk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-csk58_openstack(90346ebb-804f-449b-a13b-dab0fe49ffa2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 07:10:21 crc kubenswrapper[4809]: E1124 07:10:21.961498 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-csk58" podUID="90346ebb-804f-449b-a13b-dab0fe49ffa2" Nov 24 07:10:21 crc kubenswrapper[4809]: E1124 07:10:21.979238 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 07:10:21 crc kubenswrapper[4809]: E1124 07:10:21.979472 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dtfdq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-dbv92_openstack(8430d1e7-0ad2-4e98-aec8-e4fb4da197fe): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 07:10:21 crc kubenswrapper[4809]: E1124 07:10:21.981660 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-dbv92" podUID="8430d1e7-0ad2-4e98-aec8-e4fb4da197fe" Nov 24 07:10:22 crc kubenswrapper[4809]: I1124 07:10:22.633568 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f91f6ab4-0c7c-4002-aad9-dcd408c9000e","Type":"ContainerStarted","Data":"d661a253ac318aced47d491b5aefdb4281e8dec6c3f04fff943e9ad0ef6cbf97"} Nov 24 07:10:22 crc kubenswrapper[4809]: I1124 07:10:22.767526 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dzjmv"] Nov 24 07:10:22 crc kubenswrapper[4809]: I1124 07:10:22.778181 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:10:22 crc kubenswrapper[4809]: I1124 07:10:22.783240 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vnpwc"] Nov 24 07:10:22 crc kubenswrapper[4809]: I1124 07:10:22.794727 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 07:10:22 crc kubenswrapper[4809]: I1124 07:10:22.803016 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 07:10:22 crc kubenswrapper[4809]: I1124 07:10:22.811201 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 07:10:22 crc kubenswrapper[4809]: I1124 07:10:22.819276 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 07:10:22 crc kubenswrapper[4809]: I1124 07:10:22.824501 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jslst"] Nov 24 07:10:22 crc kubenswrapper[4809]: W1124 07:10:22.856659 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod652f6c1f_9038_4993_a8c2_fb3e9a6a822f.slice/crio-e2f10f5bdc588121fc75edc1afa0205cbaa3121f567678594da10a3eca0b0335 WatchSource:0}: Error finding container e2f10f5bdc588121fc75edc1afa0205cbaa3121f567678594da10a3eca0b0335: Status 404 returned error can't find the container with id e2f10f5bdc588121fc75edc1afa0205cbaa3121f567678594da10a3eca0b0335 Nov 24 07:10:22 crc kubenswrapper[4809]: W1124 07:10:22.862008 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcaea018d_04b5_4d20_8161_9a9a87f67583.slice/crio-16a2a9107fad9be21cd6095140d6ca6e20bb121ba7b6c3b3e80762f56bfabb0e WatchSource:0}: Error finding container 16a2a9107fad9be21cd6095140d6ca6e20bb121ba7b6c3b3e80762f56bfabb0e: Status 404 returned error can't find the container with id 16a2a9107fad9be21cd6095140d6ca6e20bb121ba7b6c3b3e80762f56bfabb0e Nov 24 07:10:22 crc kubenswrapper[4809]: I1124 07:10:22.981117 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-dbv92" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.104620 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-drz48"] Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.134688 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8430d1e7-0ad2-4e98-aec8-e4fb4da197fe-dns-svc\") pod \"8430d1e7-0ad2-4e98-aec8-e4fb4da197fe\" (UID: \"8430d1e7-0ad2-4e98-aec8-e4fb4da197fe\") " Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.135005 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8430d1e7-0ad2-4e98-aec8-e4fb4da197fe-config\") pod \"8430d1e7-0ad2-4e98-aec8-e4fb4da197fe\" (UID: \"8430d1e7-0ad2-4e98-aec8-e4fb4da197fe\") " Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.135127 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtfdq\" (UniqueName: \"kubernetes.io/projected/8430d1e7-0ad2-4e98-aec8-e4fb4da197fe-kube-api-access-dtfdq\") pod \"8430d1e7-0ad2-4e98-aec8-e4fb4da197fe\" (UID: \"8430d1e7-0ad2-4e98-aec8-e4fb4da197fe\") " Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.136509 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8430d1e7-0ad2-4e98-aec8-e4fb4da197fe-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8430d1e7-0ad2-4e98-aec8-e4fb4da197fe" (UID: "8430d1e7-0ad2-4e98-aec8-e4fb4da197fe"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.136584 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8430d1e7-0ad2-4e98-aec8-e4fb4da197fe-config" (OuterVolumeSpecName: "config") pod "8430d1e7-0ad2-4e98-aec8-e4fb4da197fe" (UID: "8430d1e7-0ad2-4e98-aec8-e4fb4da197fe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.187157 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8430d1e7-0ad2-4e98-aec8-e4fb4da197fe-kube-api-access-dtfdq" (OuterVolumeSpecName: "kube-api-access-dtfdq") pod "8430d1e7-0ad2-4e98-aec8-e4fb4da197fe" (UID: "8430d1e7-0ad2-4e98-aec8-e4fb4da197fe"). InnerVolumeSpecName "kube-api-access-dtfdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.218079 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-mp9kw"] Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.219142 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-mp9kw" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.225016 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.234226 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-csk58" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.234450 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.236933 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8430d1e7-0ad2-4e98-aec8-e4fb4da197fe-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.237074 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtfdq\" (UniqueName: \"kubernetes.io/projected/8430d1e7-0ad2-4e98-aec8-e4fb4da197fe-kube-api-access-dtfdq\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.237132 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8430d1e7-0ad2-4e98-aec8-e4fb4da197fe-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.239725 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-mp9kw"] Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.338531 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90346ebb-804f-449b-a13b-dab0fe49ffa2-config\") pod \"90346ebb-804f-449b-a13b-dab0fe49ffa2\" (UID: \"90346ebb-804f-449b-a13b-dab0fe49ffa2\") " Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.338722 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcnlk\" (UniqueName: \"kubernetes.io/projected/90346ebb-804f-449b-a13b-dab0fe49ffa2-kube-api-access-wcnlk\") pod \"90346ebb-804f-449b-a13b-dab0fe49ffa2\" (UID: \"90346ebb-804f-449b-a13b-dab0fe49ffa2\") " Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.339164 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90346ebb-804f-449b-a13b-dab0fe49ffa2-config" (OuterVolumeSpecName: "config") pod "90346ebb-804f-449b-a13b-dab0fe49ffa2" (UID: "90346ebb-804f-449b-a13b-dab0fe49ffa2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.339887 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/781916ac-bfd7-4ce4-85c0-1e938fae3103-ovs-rundir\") pod \"ovn-controller-metrics-mp9kw\" (UID: \"781916ac-bfd7-4ce4-85c0-1e938fae3103\") " pod="openstack/ovn-controller-metrics-mp9kw" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.340023 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-886fv\" (UniqueName: \"kubernetes.io/projected/781916ac-bfd7-4ce4-85c0-1e938fae3103-kube-api-access-886fv\") pod \"ovn-controller-metrics-mp9kw\" (UID: \"781916ac-bfd7-4ce4-85c0-1e938fae3103\") " pod="openstack/ovn-controller-metrics-mp9kw" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.340079 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/781916ac-bfd7-4ce4-85c0-1e938fae3103-ovn-rundir\") pod \"ovn-controller-metrics-mp9kw\" (UID: \"781916ac-bfd7-4ce4-85c0-1e938fae3103\") " pod="openstack/ovn-controller-metrics-mp9kw" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.340101 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/781916ac-bfd7-4ce4-85c0-1e938fae3103-combined-ca-bundle\") pod \"ovn-controller-metrics-mp9kw\" (UID: \"781916ac-bfd7-4ce4-85c0-1e938fae3103\") " pod="openstack/ovn-controller-metrics-mp9kw" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.340140 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/781916ac-bfd7-4ce4-85c0-1e938fae3103-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-mp9kw\" (UID: \"781916ac-bfd7-4ce4-85c0-1e938fae3103\") " pod="openstack/ovn-controller-metrics-mp9kw" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.340195 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/781916ac-bfd7-4ce4-85c0-1e938fae3103-config\") pod \"ovn-controller-metrics-mp9kw\" (UID: \"781916ac-bfd7-4ce4-85c0-1e938fae3103\") " pod="openstack/ovn-controller-metrics-mp9kw" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.340256 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90346ebb-804f-449b-a13b-dab0fe49ffa2-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.342023 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90346ebb-804f-449b-a13b-dab0fe49ffa2-kube-api-access-wcnlk" (OuterVolumeSpecName: "kube-api-access-wcnlk") pod "90346ebb-804f-449b-a13b-dab0fe49ffa2" (UID: "90346ebb-804f-449b-a13b-dab0fe49ffa2"). InnerVolumeSpecName "kube-api-access-wcnlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.441261 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/781916ac-bfd7-4ce4-85c0-1e938fae3103-config\") pod \"ovn-controller-metrics-mp9kw\" (UID: \"781916ac-bfd7-4ce4-85c0-1e938fae3103\") " pod="openstack/ovn-controller-metrics-mp9kw" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.441349 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/781916ac-bfd7-4ce4-85c0-1e938fae3103-ovs-rundir\") pod \"ovn-controller-metrics-mp9kw\" (UID: \"781916ac-bfd7-4ce4-85c0-1e938fae3103\") " pod="openstack/ovn-controller-metrics-mp9kw" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.441404 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-886fv\" (UniqueName: \"kubernetes.io/projected/781916ac-bfd7-4ce4-85c0-1e938fae3103-kube-api-access-886fv\") pod \"ovn-controller-metrics-mp9kw\" (UID: \"781916ac-bfd7-4ce4-85c0-1e938fae3103\") " pod="openstack/ovn-controller-metrics-mp9kw" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.441457 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/781916ac-bfd7-4ce4-85c0-1e938fae3103-ovn-rundir\") pod \"ovn-controller-metrics-mp9kw\" (UID: \"781916ac-bfd7-4ce4-85c0-1e938fae3103\") " pod="openstack/ovn-controller-metrics-mp9kw" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.441486 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/781916ac-bfd7-4ce4-85c0-1e938fae3103-combined-ca-bundle\") pod \"ovn-controller-metrics-mp9kw\" (UID: \"781916ac-bfd7-4ce4-85c0-1e938fae3103\") " pod="openstack/ovn-controller-metrics-mp9kw" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.441525 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/781916ac-bfd7-4ce4-85c0-1e938fae3103-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-mp9kw\" (UID: \"781916ac-bfd7-4ce4-85c0-1e938fae3103\") " pod="openstack/ovn-controller-metrics-mp9kw" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.441599 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcnlk\" (UniqueName: \"kubernetes.io/projected/90346ebb-804f-449b-a13b-dab0fe49ffa2-kube-api-access-wcnlk\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.441955 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/781916ac-bfd7-4ce4-85c0-1e938fae3103-config\") pod \"ovn-controller-metrics-mp9kw\" (UID: \"781916ac-bfd7-4ce4-85c0-1e938fae3103\") " pod="openstack/ovn-controller-metrics-mp9kw" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.442272 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/781916ac-bfd7-4ce4-85c0-1e938fae3103-ovn-rundir\") pod \"ovn-controller-metrics-mp9kw\" (UID: \"781916ac-bfd7-4ce4-85c0-1e938fae3103\") " pod="openstack/ovn-controller-metrics-mp9kw" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.442471 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/781916ac-bfd7-4ce4-85c0-1e938fae3103-ovs-rundir\") pod \"ovn-controller-metrics-mp9kw\" (UID: \"781916ac-bfd7-4ce4-85c0-1e938fae3103\") " pod="openstack/ovn-controller-metrics-mp9kw" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.447068 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/781916ac-bfd7-4ce4-85c0-1e938fae3103-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-mp9kw\" (UID: \"781916ac-bfd7-4ce4-85c0-1e938fae3103\") " pod="openstack/ovn-controller-metrics-mp9kw" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.448044 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/781916ac-bfd7-4ce4-85c0-1e938fae3103-combined-ca-bundle\") pod \"ovn-controller-metrics-mp9kw\" (UID: \"781916ac-bfd7-4ce4-85c0-1e938fae3103\") " pod="openstack/ovn-controller-metrics-mp9kw" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.457203 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-886fv\" (UniqueName: \"kubernetes.io/projected/781916ac-bfd7-4ce4-85c0-1e938fae3103-kube-api-access-886fv\") pod \"ovn-controller-metrics-mp9kw\" (UID: \"781916ac-bfd7-4ce4-85c0-1e938fae3103\") " pod="openstack/ovn-controller-metrics-mp9kw" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.537569 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-mp9kw" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.643267 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"652f6c1f-9038-4993-a8c2-fb3e9a6a822f","Type":"ContainerStarted","Data":"e2f10f5bdc588121fc75edc1afa0205cbaa3121f567678594da10a3eca0b0335"} Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.644927 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-dbv92" event={"ID":"8430d1e7-0ad2-4e98-aec8-e4fb4da197fe","Type":"ContainerDied","Data":"cfaaf47c03c98bd962e63d1cf23909282b1c619bada839803bbea8ca98bcd1f4"} Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.645004 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-dbv92" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.653339 4809 generic.go:334] "Generic (PLEG): container finished" podID="dbcae289-c075-4ecf-997d-64f15f6a8fbf" containerID="a48704a4ca4f7583baecf1a43b42c7171cae6ebf32ed2e485faf88edad8a2ba8" exitCode=0 Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.653413 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-dzjmv" event={"ID":"dbcae289-c075-4ecf-997d-64f15f6a8fbf","Type":"ContainerDied","Data":"a48704a4ca4f7583baecf1a43b42c7171cae6ebf32ed2e485faf88edad8a2ba8"} Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.653444 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-dzjmv" event={"ID":"dbcae289-c075-4ecf-997d-64f15f6a8fbf","Type":"ContainerStarted","Data":"ab3580c27c5debb7ff9f8a5fda045b015aba7d61d58064d408d3e763d3167c19"} Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.687629 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88","Type":"ContainerStarted","Data":"771058bc9399007899cbc389fa035a3f41f6239ff2ff4a67068717f5c28b6371"} Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.698063 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-jslst" event={"ID":"1041a99c-ae53-4ec0-9f44-bc33025d8566","Type":"ContainerStarted","Data":"f093eaab55c0e1c0654d590f9bfb2de554839c98cfa340436451cb0da6357cf6"} Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.715948 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-csk58" event={"ID":"90346ebb-804f-449b-a13b-dab0fe49ffa2","Type":"ContainerDied","Data":"74f1cab12fccf0ce837d8f52b995befab27b485313681a938049d69100aa266a"} Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.716072 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-csk58" Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.730349 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"caea018d-04b5-4d20-8161-9a9a87f67583","Type":"ContainerStarted","Data":"16a2a9107fad9be21cd6095140d6ca6e20bb121ba7b6c3b3e80762f56bfabb0e"} Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.741889 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7d0a537c-b1f0-4908-a604-424ea6a58438","Type":"ContainerStarted","Data":"64b042e315d4bc4990612df074423d1bf871c1b70bca6735d37406e07036dcb3"} Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.783249 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-dbv92"] Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.787652 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-dbv92"] Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.790331 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"ced578a4-f9b1-4d1f-978f-beda4b33bc59","Type":"ContainerStarted","Data":"4c9b0ede8a8c80b55c0acf6bce45d73b0996c2d749ff78df7a3b22b287505bbd"} Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.792277 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-drz48" event={"ID":"3c2d3d10-789d-4a2d-9926-cba915a127df","Type":"ContainerStarted","Data":"773311312bdc493be2af1e1e21c73660edf458e14dbe1221a4b0870748990329"} Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.794021 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vnpwc" event={"ID":"c14ae567-41ab-455a-8756-04daf7c81677","Type":"ContainerStarted","Data":"28fcfd53ab78169e59cfbfa314d530b190e781395a8356b0424cb5168574f93e"} Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.834897 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-csk58"] Nov 24 07:10:23 crc kubenswrapper[4809]: I1124 07:10:23.843404 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-csk58"] Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:23.989135 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:23.992322 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:23.995162 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-58jjb" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:23.997531 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:23.997827 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:23.998062 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.001780 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.050263 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4be21c61-834b-4cff-bae7-ec458991ac7c-config\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.050319 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4be21c61-834b-4cff-bae7-ec458991ac7c-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.050375 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf77x\" (UniqueName: \"kubernetes.io/projected/4be21c61-834b-4cff-bae7-ec458991ac7c-kube-api-access-kf77x\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.050414 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4be21c61-834b-4cff-bae7-ec458991ac7c-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.050436 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4be21c61-834b-4cff-bae7-ec458991ac7c-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.050464 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4be21c61-834b-4cff-bae7-ec458991ac7c-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.050512 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.050528 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4be21c61-834b-4cff-bae7-ec458991ac7c-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.151740 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4be21c61-834b-4cff-bae7-ec458991ac7c-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.151800 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4be21c61-834b-4cff-bae7-ec458991ac7c-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.151841 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4be21c61-834b-4cff-bae7-ec458991ac7c-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.151910 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.151931 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4be21c61-834b-4cff-bae7-ec458991ac7c-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.151985 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4be21c61-834b-4cff-bae7-ec458991ac7c-config\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.152022 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4be21c61-834b-4cff-bae7-ec458991ac7c-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.152069 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf77x\" (UniqueName: \"kubernetes.io/projected/4be21c61-834b-4cff-bae7-ec458991ac7c-kube-api-access-kf77x\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.153160 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4be21c61-834b-4cff-bae7-ec458991ac7c-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.153751 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.154472 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4be21c61-834b-4cff-bae7-ec458991ac7c-config\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.154918 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4be21c61-834b-4cff-bae7-ec458991ac7c-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.158909 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4be21c61-834b-4cff-bae7-ec458991ac7c-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.167433 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4be21c61-834b-4cff-bae7-ec458991ac7c-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.182990 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf77x\" (UniqueName: \"kubernetes.io/projected/4be21c61-834b-4cff-bae7-ec458991ac7c-kube-api-access-kf77x\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.189169 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4be21c61-834b-4cff-bae7-ec458991ac7c-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.213914 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"4be21c61-834b-4cff-bae7-ec458991ac7c\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.317122 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.726595 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.728256 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.730794 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.731142 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.731440 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-lcfv7" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.731611 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.760831 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.762102 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.762176 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.762255 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.762297 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.762358 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhd84\" (UniqueName: \"kubernetes.io/projected/ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe-kube-api-access-zhd84\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.762390 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe-config\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.762501 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.762565 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.836019 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-mp9kw"] Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.864883 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.865207 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.865403 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.865542 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.865658 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.865799 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhd84\" (UniqueName: \"kubernetes.io/projected/ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe-kube-api-access-zhd84\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.865951 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.865844 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.866248 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe-config\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.866437 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.866751 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe-config\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.867652 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.869831 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.870478 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.876925 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.882716 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhd84\" (UniqueName: \"kubernetes.io/projected/ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe-kube-api-access-zhd84\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.887278 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.903522 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8430d1e7-0ad2-4e98-aec8-e4fb4da197fe" path="/var/lib/kubelet/pods/8430d1e7-0ad2-4e98-aec8-e4fb4da197fe/volumes" Nov 24 07:10:24 crc kubenswrapper[4809]: I1124 07:10:24.903876 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90346ebb-804f-449b-a13b-dab0fe49ffa2" path="/var/lib/kubelet/pods/90346ebb-804f-449b-a13b-dab0fe49ffa2/volumes" Nov 24 07:10:25 crc kubenswrapper[4809]: I1124 07:10:25.055250 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:25 crc kubenswrapper[4809]: I1124 07:10:25.818043 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-mp9kw" event={"ID":"781916ac-bfd7-4ce4-85c0-1e938fae3103","Type":"ContainerStarted","Data":"047a78c33aa7f93aa792309e5d5b38bdd904da81cc9c804d955487a46dae3674"} Nov 24 07:10:30 crc kubenswrapper[4809]: I1124 07:10:30.003160 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 07:10:31 crc kubenswrapper[4809]: I1124 07:10:31.622325 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 07:10:31 crc kubenswrapper[4809]: I1124 07:10:31.860226 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4be21c61-834b-4cff-bae7-ec458991ac7c","Type":"ContainerStarted","Data":"1f577f657bf8964e6fda6033bfe487f99bdba7078124de9b333ee3910f1ed1dd"} Nov 24 07:10:31 crc kubenswrapper[4809]: W1124 07:10:31.929791 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce74b3ec_a6ef_4fc4_a328_6a828b8f9afe.slice/crio-595ed83087da89266f8b3572f982aa210816cd5e7e0e0d866a460087ed73f9ed WatchSource:0}: Error finding container 595ed83087da89266f8b3572f982aa210816cd5e7e0e0d866a460087ed73f9ed: Status 404 returned error can't find the container with id 595ed83087da89266f8b3572f982aa210816cd5e7e0e0d866a460087ed73f9ed Nov 24 07:10:32 crc kubenswrapper[4809]: I1124 07:10:32.872502 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-dzjmv" event={"ID":"dbcae289-c075-4ecf-997d-64f15f6a8fbf","Type":"ContainerStarted","Data":"2a05e878faf4425f80c5248725695bf049075683ada3d99dbb9782cd565fb869"} Nov 24 07:10:32 crc kubenswrapper[4809]: I1124 07:10:32.872792 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-dzjmv" Nov 24 07:10:32 crc kubenswrapper[4809]: I1124 07:10:32.874624 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88","Type":"ContainerStarted","Data":"bad9c0cea6d6f216eeac6cce3ab3f5869c9cc3f5e7ff36be4c6a7dfc605f4baa"} Nov 24 07:10:32 crc kubenswrapper[4809]: I1124 07:10:32.877139 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe","Type":"ContainerStarted","Data":"595ed83087da89266f8b3572f982aa210816cd5e7e0e0d866a460087ed73f9ed"} Nov 24 07:10:32 crc kubenswrapper[4809]: I1124 07:10:32.880759 4809 generic.go:334] "Generic (PLEG): container finished" podID="3c2d3d10-789d-4a2d-9926-cba915a127df" containerID="76af534f5884200590ca806c34604497ad2e5d943e76f77a0f437f19722fe760" exitCode=0 Nov 24 07:10:32 crc kubenswrapper[4809]: I1124 07:10:32.880803 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-drz48" event={"ID":"3c2d3d10-789d-4a2d-9926-cba915a127df","Type":"ContainerDied","Data":"76af534f5884200590ca806c34604497ad2e5d943e76f77a0f437f19722fe760"} Nov 24 07:10:32 crc kubenswrapper[4809]: I1124 07:10:32.883040 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vnpwc" event={"ID":"c14ae567-41ab-455a-8756-04daf7c81677","Type":"ContainerStarted","Data":"708529e11e0823adc87a292e989a17015a4e96af54c1c21c814c2d0c828cad8a"} Nov 24 07:10:32 crc kubenswrapper[4809]: I1124 07:10:32.883439 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-vnpwc" Nov 24 07:10:32 crc kubenswrapper[4809]: I1124 07:10:32.884884 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"ced578a4-f9b1-4d1f-978f-beda4b33bc59","Type":"ContainerStarted","Data":"2e6eec33794af4819e77bde578f741b8ea045882b3f74bd0ae0f3ad51f470910"} Nov 24 07:10:32 crc kubenswrapper[4809]: I1124 07:10:32.885289 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 24 07:10:32 crc kubenswrapper[4809]: I1124 07:10:32.886794 4809 generic.go:334] "Generic (PLEG): container finished" podID="1041a99c-ae53-4ec0-9f44-bc33025d8566" containerID="ffb66ce78a9d5ff4e382b7d6f2521b369cabf06df3203330121999e0e7b2043d" exitCode=0 Nov 24 07:10:32 crc kubenswrapper[4809]: I1124 07:10:32.886855 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-jslst" event={"ID":"1041a99c-ae53-4ec0-9f44-bc33025d8566","Type":"ContainerDied","Data":"ffb66ce78a9d5ff4e382b7d6f2521b369cabf06df3203330121999e0e7b2043d"} Nov 24 07:10:32 crc kubenswrapper[4809]: I1124 07:10:32.889181 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"caea018d-04b5-4d20-8161-9a9a87f67583","Type":"ContainerStarted","Data":"998983f443355bce0271f64a5a867259f1eed9c32e7a7ce2a4db86f97435ea0d"} Nov 24 07:10:32 crc kubenswrapper[4809]: I1124 07:10:32.895028 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-dzjmv" podStartSLOduration=21.388306594 podStartE2EDuration="21.89500511s" podCreationTimestamp="2025-11-24 07:10:11 +0000 UTC" firstStartedPulling="2025-11-24 07:10:22.855583704 +0000 UTC m=+922.756175409" lastFinishedPulling="2025-11-24 07:10:23.36228221 +0000 UTC m=+923.262873925" observedRunningTime="2025-11-24 07:10:32.892479694 +0000 UTC m=+932.793071399" watchObservedRunningTime="2025-11-24 07:10:32.89500511 +0000 UTC m=+932.795596815" Nov 24 07:10:32 crc kubenswrapper[4809]: I1124 07:10:32.903720 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 07:10:32 crc kubenswrapper[4809]: I1124 07:10:32.903772 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"652f6c1f-9038-4993-a8c2-fb3e9a6a822f","Type":"ContainerStarted","Data":"464ead71dc25f530e7baac07291d1efa1b3c50cbfa75f7230f141ec025cefb49"} Nov 24 07:10:32 crc kubenswrapper[4809]: I1124 07:10:32.923017 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=11.34350677 podStartE2EDuration="17.923000128s" podCreationTimestamp="2025-11-24 07:10:15 +0000 UTC" firstStartedPulling="2025-11-24 07:10:22.871512597 +0000 UTC m=+922.772104302" lastFinishedPulling="2025-11-24 07:10:29.451005955 +0000 UTC m=+929.351597660" observedRunningTime="2025-11-24 07:10:32.920231136 +0000 UTC m=+932.820822851" watchObservedRunningTime="2025-11-24 07:10:32.923000128 +0000 UTC m=+932.823591833" Nov 24 07:10:32 crc kubenswrapper[4809]: I1124 07:10:32.975707 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-vnpwc" podStartSLOduration=4.065270943 podStartE2EDuration="12.975687179s" podCreationTimestamp="2025-11-24 07:10:20 +0000 UTC" firstStartedPulling="2025-11-24 07:10:22.871719073 +0000 UTC m=+922.772310778" lastFinishedPulling="2025-11-24 07:10:31.782135309 +0000 UTC m=+931.682727014" observedRunningTime="2025-11-24 07:10:32.973658536 +0000 UTC m=+932.874250341" watchObservedRunningTime="2025-11-24 07:10:32.975687179 +0000 UTC m=+932.876278884" Nov 24 07:10:33 crc kubenswrapper[4809]: E1124 07:10:33.040862 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c2d3d10_789d_4a2d_9926_cba915a127df.slice/crio-conmon-76af534f5884200590ca806c34604497ad2e5d943e76f77a0f437f19722fe760.scope\": RecentStats: unable to find data in memory cache]" Nov 24 07:10:33 crc kubenswrapper[4809]: I1124 07:10:33.904198 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe","Type":"ContainerStarted","Data":"42f17f37c598b1a7f54eb3d91f766b7f848dc78b3a70200b60b3ace27ad74904"} Nov 24 07:10:33 crc kubenswrapper[4809]: I1124 07:10:33.906302 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f91f6ab4-0c7c-4002-aad9-dcd408c9000e","Type":"ContainerStarted","Data":"ae51b70d41f1cc092e89e2f506b0d859bd045728a8e37fb6553593c1785b0da8"} Nov 24 07:10:33 crc kubenswrapper[4809]: I1124 07:10:33.908808 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4be21c61-834b-4cff-bae7-ec458991ac7c","Type":"ContainerStarted","Data":"cd2b8368df9ff476a2022fee766a9c3252a2ab3d703c78e09730d4440502545f"} Nov 24 07:10:33 crc kubenswrapper[4809]: I1124 07:10:33.908856 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4be21c61-834b-4cff-bae7-ec458991ac7c","Type":"ContainerStarted","Data":"bac5ee7d47e9b36e57bbd56a2342d3bf9a4739bff49eb37c516bea0f39e181a6"} Nov 24 07:10:33 crc kubenswrapper[4809]: I1124 07:10:33.912715 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-drz48" event={"ID":"3c2d3d10-789d-4a2d-9926-cba915a127df","Type":"ContainerStarted","Data":"dd7c22db2747e942390320a8f8b784d7a755b1f435585f18137f2e87bd31b236"} Nov 24 07:10:33 crc kubenswrapper[4809]: I1124 07:10:33.916402 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7d0a537c-b1f0-4908-a604-424ea6a58438","Type":"ContainerStarted","Data":"3e28454c296359bd8a170cd1fd9c49436203c71b0aad0ea763826736402a4456"} Nov 24 07:10:33 crc kubenswrapper[4809]: I1124 07:10:33.922002 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-jslst" event={"ID":"1041a99c-ae53-4ec0-9f44-bc33025d8566","Type":"ContainerStarted","Data":"56854a1b91dbe7c71523587c6eb879cba8e418f902ec558ee87a653f82cf4d16"} Nov 24 07:10:33 crc kubenswrapper[4809]: I1124 07:10:33.922125 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-jslst" Nov 24 07:10:33 crc kubenswrapper[4809]: I1124 07:10:33.925234 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-mp9kw" event={"ID":"781916ac-bfd7-4ce4-85c0-1e938fae3103","Type":"ContainerStarted","Data":"ff28913c9dbc1bcacf9ad69614851a09e5a1a88c81ec2806123b14f05f68dc58"} Nov 24 07:10:33 crc kubenswrapper[4809]: I1124 07:10:33.941409 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=8.766107299 podStartE2EDuration="17.941392913s" podCreationTimestamp="2025-11-24 07:10:16 +0000 UTC" firstStartedPulling="2025-11-24 07:10:22.861671702 +0000 UTC m=+922.762263407" lastFinishedPulling="2025-11-24 07:10:32.036957316 +0000 UTC m=+931.937549021" observedRunningTime="2025-11-24 07:10:33.050819923 +0000 UTC m=+932.951411638" watchObservedRunningTime="2025-11-24 07:10:33.941392913 +0000 UTC m=+933.841984618" Nov 24 07:10:33 crc kubenswrapper[4809]: I1124 07:10:33.956249 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-jslst" podStartSLOduration=22.060319465 podStartE2EDuration="23.956231669s" podCreationTimestamp="2025-11-24 07:10:10 +0000 UTC" firstStartedPulling="2025-11-24 07:10:22.872534585 +0000 UTC m=+922.773126290" lastFinishedPulling="2025-11-24 07:10:24.768446799 +0000 UTC m=+924.669038494" observedRunningTime="2025-11-24 07:10:33.955917451 +0000 UTC m=+933.856509156" watchObservedRunningTime="2025-11-24 07:10:33.956231669 +0000 UTC m=+933.856823374" Nov 24 07:10:33 crc kubenswrapper[4809]: I1124 07:10:33.996398 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=10.544911776 podStartE2EDuration="11.996381343s" podCreationTimestamp="2025-11-24 07:10:22 +0000 UTC" firstStartedPulling="2025-11-24 07:10:31.171629412 +0000 UTC m=+931.072221117" lastFinishedPulling="2025-11-24 07:10:32.623098979 +0000 UTC m=+932.523690684" observedRunningTime="2025-11-24 07:10:33.9951248 +0000 UTC m=+933.895716505" watchObservedRunningTime="2025-11-24 07:10:33.996381343 +0000 UTC m=+933.896973048" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.012318 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-mp9kw" podStartSLOduration=4.569682328 podStartE2EDuration="11.012303777s" podCreationTimestamp="2025-11-24 07:10:23 +0000 UTC" firstStartedPulling="2025-11-24 07:10:25.598693121 +0000 UTC m=+925.499284826" lastFinishedPulling="2025-11-24 07:10:32.04131453 +0000 UTC m=+931.941906275" observedRunningTime="2025-11-24 07:10:34.009275048 +0000 UTC m=+933.909866773" watchObservedRunningTime="2025-11-24 07:10:34.012303777 +0000 UTC m=+933.912895482" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.318315 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.391211 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dzjmv"] Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.427849 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-72s5m"] Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.429032 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.430870 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.438512 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-72s5m"] Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.521429 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4701e8a3-c281-4309-8998-a1dd6b99daca-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-72s5m\" (UID: \"4701e8a3-c281-4309-8998-a1dd6b99daca\") " pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.521789 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4701e8a3-c281-4309-8998-a1dd6b99daca-config\") pod \"dnsmasq-dns-7fd796d7df-72s5m\" (UID: \"4701e8a3-c281-4309-8998-a1dd6b99daca\") " pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.521936 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4701e8a3-c281-4309-8998-a1dd6b99daca-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-72s5m\" (UID: \"4701e8a3-c281-4309-8998-a1dd6b99daca\") " pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.522046 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn9sp\" (UniqueName: \"kubernetes.io/projected/4701e8a3-c281-4309-8998-a1dd6b99daca-kube-api-access-cn9sp\") pod \"dnsmasq-dns-7fd796d7df-72s5m\" (UID: \"4701e8a3-c281-4309-8998-a1dd6b99daca\") " pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.538644 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jslst"] Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.566258 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-mqhjl"] Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.579691 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.587552 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.596788 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-mqhjl"] Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.625885 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4701e8a3-c281-4309-8998-a1dd6b99daca-config\") pod \"dnsmasq-dns-7fd796d7df-72s5m\" (UID: \"4701e8a3-c281-4309-8998-a1dd6b99daca\") " pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.625978 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4701e8a3-c281-4309-8998-a1dd6b99daca-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-72s5m\" (UID: \"4701e8a3-c281-4309-8998-a1dd6b99daca\") " pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.626013 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn9sp\" (UniqueName: \"kubernetes.io/projected/4701e8a3-c281-4309-8998-a1dd6b99daca-kube-api-access-cn9sp\") pod \"dnsmasq-dns-7fd796d7df-72s5m\" (UID: \"4701e8a3-c281-4309-8998-a1dd6b99daca\") " pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.626056 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4701e8a3-c281-4309-8998-a1dd6b99daca-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-72s5m\" (UID: \"4701e8a3-c281-4309-8998-a1dd6b99daca\") " pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.626848 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4701e8a3-c281-4309-8998-a1dd6b99daca-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-72s5m\" (UID: \"4701e8a3-c281-4309-8998-a1dd6b99daca\") " pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.626939 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4701e8a3-c281-4309-8998-a1dd6b99daca-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-72s5m\" (UID: \"4701e8a3-c281-4309-8998-a1dd6b99daca\") " pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.627564 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4701e8a3-c281-4309-8998-a1dd6b99daca-config\") pod \"dnsmasq-dns-7fd796d7df-72s5m\" (UID: \"4701e8a3-c281-4309-8998-a1dd6b99daca\") " pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.643745 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn9sp\" (UniqueName: \"kubernetes.io/projected/4701e8a3-c281-4309-8998-a1dd6b99daca-kube-api-access-cn9sp\") pod \"dnsmasq-dns-7fd796d7df-72s5m\" (UID: \"4701e8a3-c281-4309-8998-a1dd6b99daca\") " pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.727914 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2pfm\" (UniqueName: \"kubernetes.io/projected/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-kube-api-access-m2pfm\") pod \"dnsmasq-dns-86db49b7ff-mqhjl\" (UID: \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\") " pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.727955 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-mqhjl\" (UID: \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\") " pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.727994 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-mqhjl\" (UID: \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\") " pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.728210 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-mqhjl\" (UID: \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\") " pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.728270 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-config\") pod \"dnsmasq-dns-86db49b7ff-mqhjl\" (UID: \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\") " pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.748082 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.830361 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-mqhjl\" (UID: \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\") " pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.830423 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-config\") pod \"dnsmasq-dns-86db49b7ff-mqhjl\" (UID: \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\") " pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.830536 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2pfm\" (UniqueName: \"kubernetes.io/projected/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-kube-api-access-m2pfm\") pod \"dnsmasq-dns-86db49b7ff-mqhjl\" (UID: \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\") " pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.830558 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-mqhjl\" (UID: \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\") " pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.830588 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-mqhjl\" (UID: \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\") " pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.831616 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-mqhjl\" (UID: \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\") " pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.832310 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-mqhjl\" (UID: \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\") " pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.832982 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-config\") pod \"dnsmasq-dns-86db49b7ff-mqhjl\" (UID: \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\") " pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.834256 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-mqhjl\" (UID: \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\") " pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.862379 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2pfm\" (UniqueName: \"kubernetes.io/projected/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-kube-api-access-m2pfm\") pod \"dnsmasq-dns-86db49b7ff-mqhjl\" (UID: \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\") " pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.895171 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.931815 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe","Type":"ContainerStarted","Data":"19634330b75b2bbd295962b9bf5400c068aa9b5d5990fae46d06b696c03dae12"} Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.937833 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-drz48" event={"ID":"3c2d3d10-789d-4a2d-9926-cba915a127df","Type":"ContainerStarted","Data":"210144eeb2c0f9489fffdcde8cb32a85b73c5395aee91d3efa937eda552881b2"} Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.937880 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.938059 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-dzjmv" podUID="dbcae289-c075-4ecf-997d-64f15f6a8fbf" containerName="dnsmasq-dns" containerID="cri-o://2a05e878faf4425f80c5248725695bf049075683ada3d99dbb9782cd565fb869" gracePeriod=10 Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.941225 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:10:34 crc kubenswrapper[4809]: I1124 07:10:34.959237 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=10.548601707 podStartE2EDuration="11.959219022s" podCreationTimestamp="2025-11-24 07:10:23 +0000 UTC" firstStartedPulling="2025-11-24 07:10:31.967058588 +0000 UTC m=+931.867650293" lastFinishedPulling="2025-11-24 07:10:33.377675903 +0000 UTC m=+933.278267608" observedRunningTime="2025-11-24 07:10:34.95486753 +0000 UTC m=+934.855459245" watchObservedRunningTime="2025-11-24 07:10:34.959219022 +0000 UTC m=+934.859810727" Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.000878 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-drz48" podStartSLOduration=6.800501136 podStartE2EDuration="15.000860036s" podCreationTimestamp="2025-11-24 07:10:20 +0000 UTC" firstStartedPulling="2025-11-24 07:10:23.18729424 +0000 UTC m=+923.087885945" lastFinishedPulling="2025-11-24 07:10:31.38765314 +0000 UTC m=+931.288244845" observedRunningTime="2025-11-24 07:10:34.994202733 +0000 UTC m=+934.894794438" watchObservedRunningTime="2025-11-24 07:10:35.000860036 +0000 UTC m=+934.901451741" Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.056185 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-72s5m"] Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.059076 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:35 crc kubenswrapper[4809]: W1124 07:10:35.081075 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4701e8a3_c281_4309_8998_a1dd6b99daca.slice/crio-478b09b24c818026cfa4671ada489e1b5b3ad2f281931908bc606f1f751d4b3a WatchSource:0}: Error finding container 478b09b24c818026cfa4671ada489e1b5b3ad2f281931908bc606f1f751d4b3a: Status 404 returned error can't find the container with id 478b09b24c818026cfa4671ada489e1b5b3ad2f281931908bc606f1f751d4b3a Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.581463 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-mqhjl"] Nov 24 07:10:35 crc kubenswrapper[4809]: W1124 07:10:35.591355 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddbd972c2_5bdf_4dd8_8a71_4ae6f6f8a85e.slice/crio-fa78fd6879fad9c918416963473c6782dbdc81434fcd4046b252642a31172d25 WatchSource:0}: Error finding container fa78fd6879fad9c918416963473c6782dbdc81434fcd4046b252642a31172d25: Status 404 returned error can't find the container with id fa78fd6879fad9c918416963473c6782dbdc81434fcd4046b252642a31172d25 Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.767018 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-dzjmv" Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.846496 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nbkk\" (UniqueName: \"kubernetes.io/projected/dbcae289-c075-4ecf-997d-64f15f6a8fbf-kube-api-access-6nbkk\") pod \"dbcae289-c075-4ecf-997d-64f15f6a8fbf\" (UID: \"dbcae289-c075-4ecf-997d-64f15f6a8fbf\") " Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.846568 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dbcae289-c075-4ecf-997d-64f15f6a8fbf-dns-svc\") pod \"dbcae289-c075-4ecf-997d-64f15f6a8fbf\" (UID: \"dbcae289-c075-4ecf-997d-64f15f6a8fbf\") " Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.846608 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbcae289-c075-4ecf-997d-64f15f6a8fbf-config\") pod \"dbcae289-c075-4ecf-997d-64f15f6a8fbf\" (UID: \"dbcae289-c075-4ecf-997d-64f15f6a8fbf\") " Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.851306 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbcae289-c075-4ecf-997d-64f15f6a8fbf-kube-api-access-6nbkk" (OuterVolumeSpecName: "kube-api-access-6nbkk") pod "dbcae289-c075-4ecf-997d-64f15f6a8fbf" (UID: "dbcae289-c075-4ecf-997d-64f15f6a8fbf"). InnerVolumeSpecName "kube-api-access-6nbkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.881777 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dbcae289-c075-4ecf-997d-64f15f6a8fbf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dbcae289-c075-4ecf-997d-64f15f6a8fbf" (UID: "dbcae289-c075-4ecf-997d-64f15f6a8fbf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.896164 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dbcae289-c075-4ecf-997d-64f15f6a8fbf-config" (OuterVolumeSpecName: "config") pod "dbcae289-c075-4ecf-997d-64f15f6a8fbf" (UID: "dbcae289-c075-4ecf-997d-64f15f6a8fbf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.947816 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nbkk\" (UniqueName: \"kubernetes.io/projected/dbcae289-c075-4ecf-997d-64f15f6a8fbf-kube-api-access-6nbkk\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.947849 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dbcae289-c075-4ecf-997d-64f15f6a8fbf-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.947862 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbcae289-c075-4ecf-997d-64f15f6a8fbf-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.953581 4809 generic.go:334] "Generic (PLEG): container finished" podID="dbcae289-c075-4ecf-997d-64f15f6a8fbf" containerID="2a05e878faf4425f80c5248725695bf049075683ada3d99dbb9782cd565fb869" exitCode=0 Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.953642 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-dzjmv" Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.953632 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-dzjmv" event={"ID":"dbcae289-c075-4ecf-997d-64f15f6a8fbf","Type":"ContainerDied","Data":"2a05e878faf4425f80c5248725695bf049075683ada3d99dbb9782cd565fb869"} Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.953776 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-dzjmv" event={"ID":"dbcae289-c075-4ecf-997d-64f15f6a8fbf","Type":"ContainerDied","Data":"ab3580c27c5debb7ff9f8a5fda045b015aba7d61d58064d408d3e763d3167c19"} Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.953805 4809 scope.go:117] "RemoveContainer" containerID="2a05e878faf4425f80c5248725695bf049075683ada3d99dbb9782cd565fb869" Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.956626 4809 generic.go:334] "Generic (PLEG): container finished" podID="dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e" containerID="3618f80dfb48da6ff14b44796854bdfabe49327124842a2b566bffb6b744df7e" exitCode=0 Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.956695 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" event={"ID":"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e","Type":"ContainerDied","Data":"3618f80dfb48da6ff14b44796854bdfabe49327124842a2b566bffb6b744df7e"} Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.956720 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" event={"ID":"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e","Type":"ContainerStarted","Data":"fa78fd6879fad9c918416963473c6782dbdc81434fcd4046b252642a31172d25"} Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.959830 4809 generic.go:334] "Generic (PLEG): container finished" podID="4701e8a3-c281-4309-8998-a1dd6b99daca" containerID="ecde444da7a8a2fb8329986a86131a4dc412a484e1d952959c708bf27f472176" exitCode=0 Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.960021 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" event={"ID":"4701e8a3-c281-4309-8998-a1dd6b99daca","Type":"ContainerDied","Data":"ecde444da7a8a2fb8329986a86131a4dc412a484e1d952959c708bf27f472176"} Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.960063 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" event={"ID":"4701e8a3-c281-4309-8998-a1dd6b99daca","Type":"ContainerStarted","Data":"478b09b24c818026cfa4671ada489e1b5b3ad2f281931908bc606f1f751d4b3a"} Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.960240 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-jslst" podUID="1041a99c-ae53-4ec0-9f44-bc33025d8566" containerName="dnsmasq-dns" containerID="cri-o://56854a1b91dbe7c71523587c6eb879cba8e418f902ec558ee87a653f82cf4d16" gracePeriod=10 Nov 24 07:10:35 crc kubenswrapper[4809]: I1124 07:10:35.988277 4809 scope.go:117] "RemoveContainer" containerID="a48704a4ca4f7583baecf1a43b42c7171cae6ebf32ed2e485faf88edad8a2ba8" Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.014125 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dzjmv"] Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.016864 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dzjmv"] Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.050668 4809 scope.go:117] "RemoveContainer" containerID="2a05e878faf4425f80c5248725695bf049075683ada3d99dbb9782cd565fb869" Nov 24 07:10:36 crc kubenswrapper[4809]: E1124 07:10:36.051920 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a05e878faf4425f80c5248725695bf049075683ada3d99dbb9782cd565fb869\": container with ID starting with 2a05e878faf4425f80c5248725695bf049075683ada3d99dbb9782cd565fb869 not found: ID does not exist" containerID="2a05e878faf4425f80c5248725695bf049075683ada3d99dbb9782cd565fb869" Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.051955 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a05e878faf4425f80c5248725695bf049075683ada3d99dbb9782cd565fb869"} err="failed to get container status \"2a05e878faf4425f80c5248725695bf049075683ada3d99dbb9782cd565fb869\": rpc error: code = NotFound desc = could not find container \"2a05e878faf4425f80c5248725695bf049075683ada3d99dbb9782cd565fb869\": container with ID starting with 2a05e878faf4425f80c5248725695bf049075683ada3d99dbb9782cd565fb869 not found: ID does not exist" Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.052007 4809 scope.go:117] "RemoveContainer" containerID="a48704a4ca4f7583baecf1a43b42c7171cae6ebf32ed2e485faf88edad8a2ba8" Nov 24 07:10:36 crc kubenswrapper[4809]: E1124 07:10:36.052277 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a48704a4ca4f7583baecf1a43b42c7171cae6ebf32ed2e485faf88edad8a2ba8\": container with ID starting with a48704a4ca4f7583baecf1a43b42c7171cae6ebf32ed2e485faf88edad8a2ba8 not found: ID does not exist" containerID="a48704a4ca4f7583baecf1a43b42c7171cae6ebf32ed2e485faf88edad8a2ba8" Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.052305 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a48704a4ca4f7583baecf1a43b42c7171cae6ebf32ed2e485faf88edad8a2ba8"} err="failed to get container status \"a48704a4ca4f7583baecf1a43b42c7171cae6ebf32ed2e485faf88edad8a2ba8\": rpc error: code = NotFound desc = could not find container \"a48704a4ca4f7583baecf1a43b42c7171cae6ebf32ed2e485faf88edad8a2ba8\": container with ID starting with a48704a4ca4f7583baecf1a43b42c7171cae6ebf32ed2e485faf88edad8a2ba8 not found: ID does not exist" Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.318383 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.371459 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jslst" Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.373778 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.558018 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1041a99c-ae53-4ec0-9f44-bc33025d8566-config\") pod \"1041a99c-ae53-4ec0-9f44-bc33025d8566\" (UID: \"1041a99c-ae53-4ec0-9f44-bc33025d8566\") " Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.558315 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1041a99c-ae53-4ec0-9f44-bc33025d8566-dns-svc\") pod \"1041a99c-ae53-4ec0-9f44-bc33025d8566\" (UID: \"1041a99c-ae53-4ec0-9f44-bc33025d8566\") " Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.558678 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mgff\" (UniqueName: \"kubernetes.io/projected/1041a99c-ae53-4ec0-9f44-bc33025d8566-kube-api-access-9mgff\") pod \"1041a99c-ae53-4ec0-9f44-bc33025d8566\" (UID: \"1041a99c-ae53-4ec0-9f44-bc33025d8566\") " Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.563176 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1041a99c-ae53-4ec0-9f44-bc33025d8566-kube-api-access-9mgff" (OuterVolumeSpecName: "kube-api-access-9mgff") pod "1041a99c-ae53-4ec0-9f44-bc33025d8566" (UID: "1041a99c-ae53-4ec0-9f44-bc33025d8566"). InnerVolumeSpecName "kube-api-access-9mgff". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.589768 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1041a99c-ae53-4ec0-9f44-bc33025d8566-config" (OuterVolumeSpecName: "config") pod "1041a99c-ae53-4ec0-9f44-bc33025d8566" (UID: "1041a99c-ae53-4ec0-9f44-bc33025d8566"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.610937 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1041a99c-ae53-4ec0-9f44-bc33025d8566-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1041a99c-ae53-4ec0-9f44-bc33025d8566" (UID: "1041a99c-ae53-4ec0-9f44-bc33025d8566"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.660265 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mgff\" (UniqueName: \"kubernetes.io/projected/1041a99c-ae53-4ec0-9f44-bc33025d8566-kube-api-access-9mgff\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.660295 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1041a99c-ae53-4ec0-9f44-bc33025d8566-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.660307 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1041a99c-ae53-4ec0-9f44-bc33025d8566-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.911739 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbcae289-c075-4ecf-997d-64f15f6a8fbf" path="/var/lib/kubelet/pods/dbcae289-c075-4ecf-997d-64f15f6a8fbf/volumes" Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.972366 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" event={"ID":"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e","Type":"ContainerStarted","Data":"e2e3af7c670851a4e17e2da182ad3af767d4ca7e2fd95e7e821fbe3d4b4a1b00"} Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.972558 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.975081 4809 generic.go:334] "Generic (PLEG): container finished" podID="e5afba7d-dbf3-4aa6-863b-1d3a6d71af88" containerID="bad9c0cea6d6f216eeac6cce3ab3f5869c9cc3f5e7ff36be4c6a7dfc605f4baa" exitCode=0 Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.975169 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88","Type":"ContainerDied","Data":"bad9c0cea6d6f216eeac6cce3ab3f5869c9cc3f5e7ff36be4c6a7dfc605f4baa"} Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.978349 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" event={"ID":"4701e8a3-c281-4309-8998-a1dd6b99daca","Type":"ContainerStarted","Data":"492cea3cd36a0c187292fb819302aebcc58ee4d18d40aa861e27cbafe0bad714"} Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.979224 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.981131 4809 generic.go:334] "Generic (PLEG): container finished" podID="1041a99c-ae53-4ec0-9f44-bc33025d8566" containerID="56854a1b91dbe7c71523587c6eb879cba8e418f902ec558ee87a653f82cf4d16" exitCode=0 Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.981204 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jslst" Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.981202 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-jslst" event={"ID":"1041a99c-ae53-4ec0-9f44-bc33025d8566","Type":"ContainerDied","Data":"56854a1b91dbe7c71523587c6eb879cba8e418f902ec558ee87a653f82cf4d16"} Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.981339 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-jslst" event={"ID":"1041a99c-ae53-4ec0-9f44-bc33025d8566","Type":"ContainerDied","Data":"f093eaab55c0e1c0654d590f9bfb2de554839c98cfa340436451cb0da6357cf6"} Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.981368 4809 scope.go:117] "RemoveContainer" containerID="56854a1b91dbe7c71523587c6eb879cba8e418f902ec558ee87a653f82cf4d16" Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.984094 4809 generic.go:334] "Generic (PLEG): container finished" podID="caea018d-04b5-4d20-8161-9a9a87f67583" containerID="998983f443355bce0271f64a5a867259f1eed9c32e7a7ce2a4db86f97435ea0d" exitCode=0 Nov 24 07:10:36 crc kubenswrapper[4809]: I1124 07:10:36.984199 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"caea018d-04b5-4d20-8161-9a9a87f67583","Type":"ContainerDied","Data":"998983f443355bce0271f64a5a867259f1eed9c32e7a7ce2a4db86f97435ea0d"} Nov 24 07:10:37 crc kubenswrapper[4809]: I1124 07:10:37.011123 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" podStartSLOduration=3.011095544 podStartE2EDuration="3.011095544s" podCreationTimestamp="2025-11-24 07:10:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:10:36.996640297 +0000 UTC m=+936.897232042" watchObservedRunningTime="2025-11-24 07:10:37.011095544 +0000 UTC m=+936.911687289" Nov 24 07:10:37 crc kubenswrapper[4809]: I1124 07:10:37.012981 4809 scope.go:117] "RemoveContainer" containerID="ffb66ce78a9d5ff4e382b7d6f2521b369cabf06df3203330121999e0e7b2043d" Nov 24 07:10:37 crc kubenswrapper[4809]: I1124 07:10:37.052938 4809 scope.go:117] "RemoveContainer" containerID="56854a1b91dbe7c71523587c6eb879cba8e418f902ec558ee87a653f82cf4d16" Nov 24 07:10:37 crc kubenswrapper[4809]: E1124 07:10:37.053518 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56854a1b91dbe7c71523587c6eb879cba8e418f902ec558ee87a653f82cf4d16\": container with ID starting with 56854a1b91dbe7c71523587c6eb879cba8e418f902ec558ee87a653f82cf4d16 not found: ID does not exist" containerID="56854a1b91dbe7c71523587c6eb879cba8e418f902ec558ee87a653f82cf4d16" Nov 24 07:10:37 crc kubenswrapper[4809]: I1124 07:10:37.053571 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56854a1b91dbe7c71523587c6eb879cba8e418f902ec558ee87a653f82cf4d16"} err="failed to get container status \"56854a1b91dbe7c71523587c6eb879cba8e418f902ec558ee87a653f82cf4d16\": rpc error: code = NotFound desc = could not find container \"56854a1b91dbe7c71523587c6eb879cba8e418f902ec558ee87a653f82cf4d16\": container with ID starting with 56854a1b91dbe7c71523587c6eb879cba8e418f902ec558ee87a653f82cf4d16 not found: ID does not exist" Nov 24 07:10:37 crc kubenswrapper[4809]: I1124 07:10:37.053606 4809 scope.go:117] "RemoveContainer" containerID="ffb66ce78a9d5ff4e382b7d6f2521b369cabf06df3203330121999e0e7b2043d" Nov 24 07:10:37 crc kubenswrapper[4809]: E1124 07:10:37.054029 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffb66ce78a9d5ff4e382b7d6f2521b369cabf06df3203330121999e0e7b2043d\": container with ID starting with ffb66ce78a9d5ff4e382b7d6f2521b369cabf06df3203330121999e0e7b2043d not found: ID does not exist" containerID="ffb66ce78a9d5ff4e382b7d6f2521b369cabf06df3203330121999e0e7b2043d" Nov 24 07:10:37 crc kubenswrapper[4809]: I1124 07:10:37.054071 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffb66ce78a9d5ff4e382b7d6f2521b369cabf06df3203330121999e0e7b2043d"} err="failed to get container status \"ffb66ce78a9d5ff4e382b7d6f2521b369cabf06df3203330121999e0e7b2043d\": rpc error: code = NotFound desc = could not find container \"ffb66ce78a9d5ff4e382b7d6f2521b369cabf06df3203330121999e0e7b2043d\": container with ID starting with ffb66ce78a9d5ff4e382b7d6f2521b369cabf06df3203330121999e0e7b2043d not found: ID does not exist" Nov 24 07:10:37 crc kubenswrapper[4809]: I1124 07:10:37.055664 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:37 crc kubenswrapper[4809]: I1124 07:10:37.094805 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jslst"] Nov 24 07:10:37 crc kubenswrapper[4809]: I1124 07:10:37.103925 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jslst"] Nov 24 07:10:37 crc kubenswrapper[4809]: I1124 07:10:37.107163 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" podStartSLOduration=3.107143502 podStartE2EDuration="3.107143502s" podCreationTimestamp="2025-11-24 07:10:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:10:37.102578863 +0000 UTC m=+937.003170598" watchObservedRunningTime="2025-11-24 07:10:37.107143502 +0000 UTC m=+937.007735217" Nov 24 07:10:37 crc kubenswrapper[4809]: I1124 07:10:37.128455 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:37 crc kubenswrapper[4809]: I1124 07:10:37.297130 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 07:10:37 crc kubenswrapper[4809]: I1124 07:10:37.995763 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e5afba7d-dbf3-4aa6-863b-1d3a6d71af88","Type":"ContainerStarted","Data":"88a319a7d6a765516a18ce81ad3058990dcddef6ddbed2ae03a8374dda5ad370"} Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.000534 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"caea018d-04b5-4d20-8161-9a9a87f67583","Type":"ContainerStarted","Data":"2b13c85e7c3cb5a40b8b713f1bc247d308f8afd8124b8660b6cf0c744f3099dc"} Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.030720 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=16.446098358 podStartE2EDuration="25.0307038s" podCreationTimestamp="2025-11-24 07:10:13 +0000 UTC" firstStartedPulling="2025-11-24 07:10:22.856720613 +0000 UTC m=+922.757312318" lastFinishedPulling="2025-11-24 07:10:31.441326055 +0000 UTC m=+931.341917760" observedRunningTime="2025-11-24 07:10:38.024115748 +0000 UTC m=+937.924707453" watchObservedRunningTime="2025-11-24 07:10:38.0307038 +0000 UTC m=+937.931295505" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.058560 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=16.941889043 podStartE2EDuration="26.058532033s" podCreationTimestamp="2025-11-24 07:10:12 +0000 UTC" firstStartedPulling="2025-11-24 07:10:22.871396404 +0000 UTC m=+922.771988109" lastFinishedPulling="2025-11-24 07:10:31.988039404 +0000 UTC m=+931.888631099" observedRunningTime="2025-11-24 07:10:38.05573099 +0000 UTC m=+937.956322695" watchObservedRunningTime="2025-11-24 07:10:38.058532033 +0000 UTC m=+937.959123768" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.067504 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.091635 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.314923 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 24 07:10:38 crc kubenswrapper[4809]: E1124 07:10:38.315241 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbcae289-c075-4ecf-997d-64f15f6a8fbf" containerName="dnsmasq-dns" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.315252 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbcae289-c075-4ecf-997d-64f15f6a8fbf" containerName="dnsmasq-dns" Nov 24 07:10:38 crc kubenswrapper[4809]: E1124 07:10:38.315265 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1041a99c-ae53-4ec0-9f44-bc33025d8566" containerName="dnsmasq-dns" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.315271 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1041a99c-ae53-4ec0-9f44-bc33025d8566" containerName="dnsmasq-dns" Nov 24 07:10:38 crc kubenswrapper[4809]: E1124 07:10:38.315289 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1041a99c-ae53-4ec0-9f44-bc33025d8566" containerName="init" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.315296 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1041a99c-ae53-4ec0-9f44-bc33025d8566" containerName="init" Nov 24 07:10:38 crc kubenswrapper[4809]: E1124 07:10:38.315314 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbcae289-c075-4ecf-997d-64f15f6a8fbf" containerName="init" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.315320 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbcae289-c075-4ecf-997d-64f15f6a8fbf" containerName="init" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.315483 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1041a99c-ae53-4ec0-9f44-bc33025d8566" containerName="dnsmasq-dns" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.315493 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbcae289-c075-4ecf-997d-64f15f6a8fbf" containerName="dnsmasq-dns" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.316280 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.318512 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-8snk5" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.321849 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.321896 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.322554 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.341023 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.391233 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acc130b5-9141-4d3a-aec1-7778963967b0-config\") pod \"ovn-northd-0\" (UID: \"acc130b5-9141-4d3a-aec1-7778963967b0\") " pod="openstack/ovn-northd-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.391387 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47wl9\" (UniqueName: \"kubernetes.io/projected/acc130b5-9141-4d3a-aec1-7778963967b0-kube-api-access-47wl9\") pod \"ovn-northd-0\" (UID: \"acc130b5-9141-4d3a-aec1-7778963967b0\") " pod="openstack/ovn-northd-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.391413 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acc130b5-9141-4d3a-aec1-7778963967b0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"acc130b5-9141-4d3a-aec1-7778963967b0\") " pod="openstack/ovn-northd-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.391442 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/acc130b5-9141-4d3a-aec1-7778963967b0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"acc130b5-9141-4d3a-aec1-7778963967b0\") " pod="openstack/ovn-northd-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.391568 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/acc130b5-9141-4d3a-aec1-7778963967b0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"acc130b5-9141-4d3a-aec1-7778963967b0\") " pod="openstack/ovn-northd-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.391799 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/acc130b5-9141-4d3a-aec1-7778963967b0-scripts\") pod \"ovn-northd-0\" (UID: \"acc130b5-9141-4d3a-aec1-7778963967b0\") " pod="openstack/ovn-northd-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.391836 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/acc130b5-9141-4d3a-aec1-7778963967b0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"acc130b5-9141-4d3a-aec1-7778963967b0\") " pod="openstack/ovn-northd-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.493204 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/acc130b5-9141-4d3a-aec1-7778963967b0-scripts\") pod \"ovn-northd-0\" (UID: \"acc130b5-9141-4d3a-aec1-7778963967b0\") " pod="openstack/ovn-northd-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.493317 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/acc130b5-9141-4d3a-aec1-7778963967b0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"acc130b5-9141-4d3a-aec1-7778963967b0\") " pod="openstack/ovn-northd-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.493387 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acc130b5-9141-4d3a-aec1-7778963967b0-config\") pod \"ovn-northd-0\" (UID: \"acc130b5-9141-4d3a-aec1-7778963967b0\") " pod="openstack/ovn-northd-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.493526 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47wl9\" (UniqueName: \"kubernetes.io/projected/acc130b5-9141-4d3a-aec1-7778963967b0-kube-api-access-47wl9\") pod \"ovn-northd-0\" (UID: \"acc130b5-9141-4d3a-aec1-7778963967b0\") " pod="openstack/ovn-northd-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.493567 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acc130b5-9141-4d3a-aec1-7778963967b0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"acc130b5-9141-4d3a-aec1-7778963967b0\") " pod="openstack/ovn-northd-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.493617 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/acc130b5-9141-4d3a-aec1-7778963967b0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"acc130b5-9141-4d3a-aec1-7778963967b0\") " pod="openstack/ovn-northd-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.493659 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/acc130b5-9141-4d3a-aec1-7778963967b0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"acc130b5-9141-4d3a-aec1-7778963967b0\") " pod="openstack/ovn-northd-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.494325 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/acc130b5-9141-4d3a-aec1-7778963967b0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"acc130b5-9141-4d3a-aec1-7778963967b0\") " pod="openstack/ovn-northd-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.494365 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/acc130b5-9141-4d3a-aec1-7778963967b0-scripts\") pod \"ovn-northd-0\" (UID: \"acc130b5-9141-4d3a-aec1-7778963967b0\") " pod="openstack/ovn-northd-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.494512 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acc130b5-9141-4d3a-aec1-7778963967b0-config\") pod \"ovn-northd-0\" (UID: \"acc130b5-9141-4d3a-aec1-7778963967b0\") " pod="openstack/ovn-northd-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.499088 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/acc130b5-9141-4d3a-aec1-7778963967b0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"acc130b5-9141-4d3a-aec1-7778963967b0\") " pod="openstack/ovn-northd-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.500050 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/acc130b5-9141-4d3a-aec1-7778963967b0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"acc130b5-9141-4d3a-aec1-7778963967b0\") " pod="openstack/ovn-northd-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.500357 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acc130b5-9141-4d3a-aec1-7778963967b0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"acc130b5-9141-4d3a-aec1-7778963967b0\") " pod="openstack/ovn-northd-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.510352 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47wl9\" (UniqueName: \"kubernetes.io/projected/acc130b5-9141-4d3a-aec1-7778963967b0-kube-api-access-47wl9\") pod \"ovn-northd-0\" (UID: \"acc130b5-9141-4d3a-aec1-7778963967b0\") " pod="openstack/ovn-northd-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.631570 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 07:10:38 crc kubenswrapper[4809]: I1124 07:10:38.900332 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1041a99c-ae53-4ec0-9f44-bc33025d8566" path="/var/lib/kubelet/pods/1041a99c-ae53-4ec0-9f44-bc33025d8566/volumes" Nov 24 07:10:39 crc kubenswrapper[4809]: I1124 07:10:39.107906 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 07:10:40 crc kubenswrapper[4809]: I1124 07:10:40.023179 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"acc130b5-9141-4d3a-aec1-7778963967b0","Type":"ContainerStarted","Data":"85cc8f3e85aba56da65e504edefdeea011e3207268120c9d7e69d3d229dc225c"} Nov 24 07:10:40 crc kubenswrapper[4809]: I1124 07:10:40.712374 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 24 07:10:43 crc kubenswrapper[4809]: I1124 07:10:43.889296 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 24 07:10:43 crc kubenswrapper[4809]: I1124 07:10:43.890018 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 24 07:10:44 crc kubenswrapper[4809]: I1124 07:10:44.750184 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" Nov 24 07:10:44 crc kubenswrapper[4809]: I1124 07:10:44.905289 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" Nov 24 07:10:44 crc kubenswrapper[4809]: I1124 07:10:44.951833 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-72s5m"] Nov 24 07:10:45 crc kubenswrapper[4809]: I1124 07:10:45.068508 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" podUID="4701e8a3-c281-4309-8998-a1dd6b99daca" containerName="dnsmasq-dns" containerID="cri-o://492cea3cd36a0c187292fb819302aebcc58ee4d18d40aa861e27cbafe0bad714" gracePeriod=10 Nov 24 07:10:45 crc kubenswrapper[4809]: I1124 07:10:45.285699 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:45 crc kubenswrapper[4809]: I1124 07:10:45.285935 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:45 crc kubenswrapper[4809]: I1124 07:10:45.354102 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:45 crc kubenswrapper[4809]: I1124 07:10:45.975762 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.069709 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.076915 4809 generic.go:334] "Generic (PLEG): container finished" podID="4701e8a3-c281-4309-8998-a1dd6b99daca" containerID="492cea3cd36a0c187292fb819302aebcc58ee4d18d40aa861e27cbafe0bad714" exitCode=0 Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.076997 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" event={"ID":"4701e8a3-c281-4309-8998-a1dd6b99daca","Type":"ContainerDied","Data":"492cea3cd36a0c187292fb819302aebcc58ee4d18d40aa861e27cbafe0bad714"} Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.077038 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" event={"ID":"4701e8a3-c281-4309-8998-a1dd6b99daca","Type":"ContainerDied","Data":"478b09b24c818026cfa4671ada489e1b5b3ad2f281931908bc606f1f751d4b3a"} Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.077055 4809 scope.go:117] "RemoveContainer" containerID="492cea3cd36a0c187292fb819302aebcc58ee4d18d40aa861e27cbafe0bad714" Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.077325 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-72s5m" Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.079200 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"acc130b5-9141-4d3a-aec1-7778963967b0","Type":"ContainerStarted","Data":"adc9bbf6acc5aa23f4e21ecbac7c7924fa356b7a03a51f4da3e4b520eb6ce0a9"} Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.079229 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"acc130b5-9141-4d3a-aec1-7778963967b0","Type":"ContainerStarted","Data":"9a4453c8fc244e02be87fe95799d8c1b39f26a2e9db531780d0eeac7a3ea3a93"} Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.079352 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.113606 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.666388477 podStartE2EDuration="8.113584154s" podCreationTimestamp="2025-11-24 07:10:38 +0000 UTC" firstStartedPulling="2025-11-24 07:10:39.109030792 +0000 UTC m=+939.009622497" lastFinishedPulling="2025-11-24 07:10:45.556226469 +0000 UTC m=+945.456818174" observedRunningTime="2025-11-24 07:10:46.107698201 +0000 UTC m=+946.008289906" watchObservedRunningTime="2025-11-24 07:10:46.113584154 +0000 UTC m=+946.014175859" Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.121399 4809 scope.go:117] "RemoveContainer" containerID="ecde444da7a8a2fb8329986a86131a4dc412a484e1d952959c708bf27f472176" Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.122121 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4701e8a3-c281-4309-8998-a1dd6b99daca-ovsdbserver-nb\") pod \"4701e8a3-c281-4309-8998-a1dd6b99daca\" (UID: \"4701e8a3-c281-4309-8998-a1dd6b99daca\") " Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.122319 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4701e8a3-c281-4309-8998-a1dd6b99daca-dns-svc\") pod \"4701e8a3-c281-4309-8998-a1dd6b99daca\" (UID: \"4701e8a3-c281-4309-8998-a1dd6b99daca\") " Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.122354 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4701e8a3-c281-4309-8998-a1dd6b99daca-config\") pod \"4701e8a3-c281-4309-8998-a1dd6b99daca\" (UID: \"4701e8a3-c281-4309-8998-a1dd6b99daca\") " Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.122817 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cn9sp\" (UniqueName: \"kubernetes.io/projected/4701e8a3-c281-4309-8998-a1dd6b99daca-kube-api-access-cn9sp\") pod \"4701e8a3-c281-4309-8998-a1dd6b99daca\" (UID: \"4701e8a3-c281-4309-8998-a1dd6b99daca\") " Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.127688 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4701e8a3-c281-4309-8998-a1dd6b99daca-kube-api-access-cn9sp" (OuterVolumeSpecName: "kube-api-access-cn9sp") pod "4701e8a3-c281-4309-8998-a1dd6b99daca" (UID: "4701e8a3-c281-4309-8998-a1dd6b99daca"). InnerVolumeSpecName "kube-api-access-cn9sp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.139375 4809 scope.go:117] "RemoveContainer" containerID="492cea3cd36a0c187292fb819302aebcc58ee4d18d40aa861e27cbafe0bad714" Nov 24 07:10:46 crc kubenswrapper[4809]: E1124 07:10:46.139784 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"492cea3cd36a0c187292fb819302aebcc58ee4d18d40aa861e27cbafe0bad714\": container with ID starting with 492cea3cd36a0c187292fb819302aebcc58ee4d18d40aa861e27cbafe0bad714 not found: ID does not exist" containerID="492cea3cd36a0c187292fb819302aebcc58ee4d18d40aa861e27cbafe0bad714" Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.139821 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"492cea3cd36a0c187292fb819302aebcc58ee4d18d40aa861e27cbafe0bad714"} err="failed to get container status \"492cea3cd36a0c187292fb819302aebcc58ee4d18d40aa861e27cbafe0bad714\": rpc error: code = NotFound desc = could not find container \"492cea3cd36a0c187292fb819302aebcc58ee4d18d40aa861e27cbafe0bad714\": container with ID starting with 492cea3cd36a0c187292fb819302aebcc58ee4d18d40aa861e27cbafe0bad714 not found: ID does not exist" Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.139846 4809 scope.go:117] "RemoveContainer" containerID="ecde444da7a8a2fb8329986a86131a4dc412a484e1d952959c708bf27f472176" Nov 24 07:10:46 crc kubenswrapper[4809]: E1124 07:10:46.140140 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecde444da7a8a2fb8329986a86131a4dc412a484e1d952959c708bf27f472176\": container with ID starting with ecde444da7a8a2fb8329986a86131a4dc412a484e1d952959c708bf27f472176 not found: ID does not exist" containerID="ecde444da7a8a2fb8329986a86131a4dc412a484e1d952959c708bf27f472176" Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.140166 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecde444da7a8a2fb8329986a86131a4dc412a484e1d952959c708bf27f472176"} err="failed to get container status \"ecde444da7a8a2fb8329986a86131a4dc412a484e1d952959c708bf27f472176\": rpc error: code = NotFound desc = could not find container \"ecde444da7a8a2fb8329986a86131a4dc412a484e1d952959c708bf27f472176\": container with ID starting with ecde444da7a8a2fb8329986a86131a4dc412a484e1d952959c708bf27f472176 not found: ID does not exist" Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.159379 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4701e8a3-c281-4309-8998-a1dd6b99daca-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4701e8a3-c281-4309-8998-a1dd6b99daca" (UID: "4701e8a3-c281-4309-8998-a1dd6b99daca"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.159469 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4701e8a3-c281-4309-8998-a1dd6b99daca-config" (OuterVolumeSpecName: "config") pod "4701e8a3-c281-4309-8998-a1dd6b99daca" (UID: "4701e8a3-c281-4309-8998-a1dd6b99daca"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.160556 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.166477 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4701e8a3-c281-4309-8998-a1dd6b99daca-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4701e8a3-c281-4309-8998-a1dd6b99daca" (UID: "4701e8a3-c281-4309-8998-a1dd6b99daca"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.167634 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.225183 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4701e8a3-c281-4309-8998-a1dd6b99daca-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.225213 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4701e8a3-c281-4309-8998-a1dd6b99daca-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.225226 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4701e8a3-c281-4309-8998-a1dd6b99daca-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.225238 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cn9sp\" (UniqueName: \"kubernetes.io/projected/4701e8a3-c281-4309-8998-a1dd6b99daca-kube-api-access-cn9sp\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.404273 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-72s5m"] Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.409594 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-72s5m"] Nov 24 07:10:46 crc kubenswrapper[4809]: I1124 07:10:46.899588 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4701e8a3-c281-4309-8998-a1dd6b99daca" path="/var/lib/kubelet/pods/4701e8a3-c281-4309-8998-a1dd6b99daca/volumes" Nov 24 07:10:47 crc kubenswrapper[4809]: I1124 07:10:47.326652 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-jlbv5"] Nov 24 07:10:47 crc kubenswrapper[4809]: E1124 07:10:47.327010 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4701e8a3-c281-4309-8998-a1dd6b99daca" containerName="init" Nov 24 07:10:47 crc kubenswrapper[4809]: I1124 07:10:47.327024 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4701e8a3-c281-4309-8998-a1dd6b99daca" containerName="init" Nov 24 07:10:47 crc kubenswrapper[4809]: E1124 07:10:47.327037 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4701e8a3-c281-4309-8998-a1dd6b99daca" containerName="dnsmasq-dns" Nov 24 07:10:47 crc kubenswrapper[4809]: I1124 07:10:47.327045 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4701e8a3-c281-4309-8998-a1dd6b99daca" containerName="dnsmasq-dns" Nov 24 07:10:47 crc kubenswrapper[4809]: I1124 07:10:47.327211 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4701e8a3-c281-4309-8998-a1dd6b99daca" containerName="dnsmasq-dns" Nov 24 07:10:47 crc kubenswrapper[4809]: I1124 07:10:47.328155 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-jlbv5" Nov 24 07:10:47 crc kubenswrapper[4809]: I1124 07:10:47.361498 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-jlbv5"] Nov 24 07:10:47 crc kubenswrapper[4809]: I1124 07:10:47.441277 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-jlbv5\" (UID: \"8e6016ff-f68c-423c-bef0-a015064efaaf\") " pod="openstack/dnsmasq-dns-698758b865-jlbv5" Nov 24 07:10:47 crc kubenswrapper[4809]: I1124 07:10:47.441350 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-jlbv5\" (UID: \"8e6016ff-f68c-423c-bef0-a015064efaaf\") " pod="openstack/dnsmasq-dns-698758b865-jlbv5" Nov 24 07:10:47 crc kubenswrapper[4809]: I1124 07:10:47.441428 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-config\") pod \"dnsmasq-dns-698758b865-jlbv5\" (UID: \"8e6016ff-f68c-423c-bef0-a015064efaaf\") " pod="openstack/dnsmasq-dns-698758b865-jlbv5" Nov 24 07:10:47 crc kubenswrapper[4809]: I1124 07:10:47.441493 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-dns-svc\") pod \"dnsmasq-dns-698758b865-jlbv5\" (UID: \"8e6016ff-f68c-423c-bef0-a015064efaaf\") " pod="openstack/dnsmasq-dns-698758b865-jlbv5" Nov 24 07:10:47 crc kubenswrapper[4809]: I1124 07:10:47.441566 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrtp8\" (UniqueName: \"kubernetes.io/projected/8e6016ff-f68c-423c-bef0-a015064efaaf-kube-api-access-hrtp8\") pod \"dnsmasq-dns-698758b865-jlbv5\" (UID: \"8e6016ff-f68c-423c-bef0-a015064efaaf\") " pod="openstack/dnsmasq-dns-698758b865-jlbv5" Nov 24 07:10:47 crc kubenswrapper[4809]: I1124 07:10:47.542610 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-jlbv5\" (UID: \"8e6016ff-f68c-423c-bef0-a015064efaaf\") " pod="openstack/dnsmasq-dns-698758b865-jlbv5" Nov 24 07:10:47 crc kubenswrapper[4809]: I1124 07:10:47.542656 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-config\") pod \"dnsmasq-dns-698758b865-jlbv5\" (UID: \"8e6016ff-f68c-423c-bef0-a015064efaaf\") " pod="openstack/dnsmasq-dns-698758b865-jlbv5" Nov 24 07:10:47 crc kubenswrapper[4809]: I1124 07:10:47.542715 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-dns-svc\") pod \"dnsmasq-dns-698758b865-jlbv5\" (UID: \"8e6016ff-f68c-423c-bef0-a015064efaaf\") " pod="openstack/dnsmasq-dns-698758b865-jlbv5" Nov 24 07:10:47 crc kubenswrapper[4809]: I1124 07:10:47.542765 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrtp8\" (UniqueName: \"kubernetes.io/projected/8e6016ff-f68c-423c-bef0-a015064efaaf-kube-api-access-hrtp8\") pod \"dnsmasq-dns-698758b865-jlbv5\" (UID: \"8e6016ff-f68c-423c-bef0-a015064efaaf\") " pod="openstack/dnsmasq-dns-698758b865-jlbv5" Nov 24 07:10:47 crc kubenswrapper[4809]: I1124 07:10:47.542796 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-jlbv5\" (UID: \"8e6016ff-f68c-423c-bef0-a015064efaaf\") " pod="openstack/dnsmasq-dns-698758b865-jlbv5" Nov 24 07:10:47 crc kubenswrapper[4809]: I1124 07:10:47.543660 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-config\") pod \"dnsmasq-dns-698758b865-jlbv5\" (UID: \"8e6016ff-f68c-423c-bef0-a015064efaaf\") " pod="openstack/dnsmasq-dns-698758b865-jlbv5" Nov 24 07:10:47 crc kubenswrapper[4809]: I1124 07:10:47.543671 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-jlbv5\" (UID: \"8e6016ff-f68c-423c-bef0-a015064efaaf\") " pod="openstack/dnsmasq-dns-698758b865-jlbv5" Nov 24 07:10:47 crc kubenswrapper[4809]: I1124 07:10:47.543674 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-dns-svc\") pod \"dnsmasq-dns-698758b865-jlbv5\" (UID: \"8e6016ff-f68c-423c-bef0-a015064efaaf\") " pod="openstack/dnsmasq-dns-698758b865-jlbv5" Nov 24 07:10:47 crc kubenswrapper[4809]: I1124 07:10:47.543717 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-jlbv5\" (UID: \"8e6016ff-f68c-423c-bef0-a015064efaaf\") " pod="openstack/dnsmasq-dns-698758b865-jlbv5" Nov 24 07:10:47 crc kubenswrapper[4809]: I1124 07:10:47.559586 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrtp8\" (UniqueName: \"kubernetes.io/projected/8e6016ff-f68c-423c-bef0-a015064efaaf-kube-api-access-hrtp8\") pod \"dnsmasq-dns-698758b865-jlbv5\" (UID: \"8e6016ff-f68c-423c-bef0-a015064efaaf\") " pod="openstack/dnsmasq-dns-698758b865-jlbv5" Nov 24 07:10:47 crc kubenswrapper[4809]: I1124 07:10:47.644360 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-jlbv5" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.043848 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.044170 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.044222 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.044906 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5bbd5cb6a3a9e9de8c1c6f89b55f7b7f733fb4a270390a494ec13ae9d9f8fbca"} pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.045000 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" containerID="cri-o://5bbd5cb6a3a9e9de8c1c6f89b55f7b7f733fb4a270390a494ec13ae9d9f8fbca" gracePeriod=600 Nov 24 07:10:48 crc kubenswrapper[4809]: W1124 07:10:48.068810 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e6016ff_f68c_423c_bef0_a015064efaaf.slice/crio-cc8a404d1bdad2dd6e9ceb8c19ee5d15acd31d58e2990398305d03446eed8ffe WatchSource:0}: Error finding container cc8a404d1bdad2dd6e9ceb8c19ee5d15acd31d58e2990398305d03446eed8ffe: Status 404 returned error can't find the container with id cc8a404d1bdad2dd6e9ceb8c19ee5d15acd31d58e2990398305d03446eed8ffe Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.072002 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-jlbv5"] Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.096302 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-jlbv5" event={"ID":"8e6016ff-f68c-423c-bef0-a015064efaaf","Type":"ContainerStarted","Data":"cc8a404d1bdad2dd6e9ceb8c19ee5d15acd31d58e2990398305d03446eed8ffe"} Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.447819 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.452772 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.454609 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.454609 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-hg2wf" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.454622 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.456053 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.467494 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.559870 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-lock\") pod \"swift-storage-0\" (UID: \"53b09c3d-ced2-431c-ab5f-2c2e09c1a686\") " pod="openstack/swift-storage-0" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.559938 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-cache\") pod \"swift-storage-0\" (UID: \"53b09c3d-ced2-431c-ab5f-2c2e09c1a686\") " pod="openstack/swift-storage-0" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.560021 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"53b09c3d-ced2-431c-ab5f-2c2e09c1a686\") " pod="openstack/swift-storage-0" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.560047 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-etc-swift\") pod \"swift-storage-0\" (UID: \"53b09c3d-ced2-431c-ab5f-2c2e09c1a686\") " pod="openstack/swift-storage-0" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.560296 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m46c\" (UniqueName: \"kubernetes.io/projected/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-kube-api-access-4m46c\") pod \"swift-storage-0\" (UID: \"53b09c3d-ced2-431c-ab5f-2c2e09c1a686\") " pod="openstack/swift-storage-0" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.661372 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-cache\") pod \"swift-storage-0\" (UID: \"53b09c3d-ced2-431c-ab5f-2c2e09c1a686\") " pod="openstack/swift-storage-0" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.661738 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"53b09c3d-ced2-431c-ab5f-2c2e09c1a686\") " pod="openstack/swift-storage-0" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.661765 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-etc-swift\") pod \"swift-storage-0\" (UID: \"53b09c3d-ced2-431c-ab5f-2c2e09c1a686\") " pod="openstack/swift-storage-0" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.661833 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m46c\" (UniqueName: \"kubernetes.io/projected/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-kube-api-access-4m46c\") pod \"swift-storage-0\" (UID: \"53b09c3d-ced2-431c-ab5f-2c2e09c1a686\") " pod="openstack/swift-storage-0" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.661855 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-lock\") pod \"swift-storage-0\" (UID: \"53b09c3d-ced2-431c-ab5f-2c2e09c1a686\") " pod="openstack/swift-storage-0" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.661854 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-cache\") pod \"swift-storage-0\" (UID: \"53b09c3d-ced2-431c-ab5f-2c2e09c1a686\") " pod="openstack/swift-storage-0" Nov 24 07:10:48 crc kubenswrapper[4809]: E1124 07:10:48.661955 4809 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 07:10:48 crc kubenswrapper[4809]: E1124 07:10:48.661983 4809 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 07:10:48 crc kubenswrapper[4809]: E1124 07:10:48.662026 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-etc-swift podName:53b09c3d-ced2-431c-ab5f-2c2e09c1a686 nodeName:}" failed. No retries permitted until 2025-11-24 07:10:49.162008468 +0000 UTC m=+949.062600173 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-etc-swift") pod "swift-storage-0" (UID: "53b09c3d-ced2-431c-ab5f-2c2e09c1a686") : configmap "swift-ring-files" not found Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.662043 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"53b09c3d-ced2-431c-ab5f-2c2e09c1a686\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/swift-storage-0" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.662428 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-lock\") pod \"swift-storage-0\" (UID: \"53b09c3d-ced2-431c-ab5f-2c2e09c1a686\") " pod="openstack/swift-storage-0" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.695880 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"53b09c3d-ced2-431c-ab5f-2c2e09c1a686\") " pod="openstack/swift-storage-0" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.700404 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m46c\" (UniqueName: \"kubernetes.io/projected/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-kube-api-access-4m46c\") pod \"swift-storage-0\" (UID: \"53b09c3d-ced2-431c-ab5f-2c2e09c1a686\") " pod="openstack/swift-storage-0" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.940204 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-8n9sj"] Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.942445 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.944409 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.944772 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.945039 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 24 07:10:48 crc kubenswrapper[4809]: I1124 07:10:48.990309 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-8n9sj"] Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.069342 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-ring-data-devices\") pod \"swift-ring-rebalance-8n9sj\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.069451 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-scripts\") pod \"swift-ring-rebalance-8n9sj\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.069471 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-combined-ca-bundle\") pod \"swift-ring-rebalance-8n9sj\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.069498 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-dispersionconf\") pod \"swift-ring-rebalance-8n9sj\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.069517 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-swiftconf\") pod \"swift-ring-rebalance-8n9sj\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.069547 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-etc-swift\") pod \"swift-ring-rebalance-8n9sj\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.069571 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42cc4\" (UniqueName: \"kubernetes.io/projected/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-kube-api-access-42cc4\") pod \"swift-ring-rebalance-8n9sj\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.108586 4809 generic.go:334] "Generic (PLEG): container finished" podID="be8050d5-1aea-453e-a06e-87702b0e856b" containerID="5bbd5cb6a3a9e9de8c1c6f89b55f7b7f733fb4a270390a494ec13ae9d9f8fbca" exitCode=0 Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.108644 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerDied","Data":"5bbd5cb6a3a9e9de8c1c6f89b55f7b7f733fb4a270390a494ec13ae9d9f8fbca"} Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.108731 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerStarted","Data":"37feb2230d7f506014799017af7463c0d7873535f5cdb2251d0334f2328d10b8"} Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.108838 4809 scope.go:117] "RemoveContainer" containerID="672cd2d0a97571eba82aabde6a902d139a7242488c7021eea3d01afc3b311c16" Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.110329 4809 generic.go:334] "Generic (PLEG): container finished" podID="8e6016ff-f68c-423c-bef0-a015064efaaf" containerID="8dbc0a02093b03837764b69c924162211844cb7df74f5c22db0eb1b2eec4d217" exitCode=0 Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.110378 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-jlbv5" event={"ID":"8e6016ff-f68c-423c-bef0-a015064efaaf","Type":"ContainerDied","Data":"8dbc0a02093b03837764b69c924162211844cb7df74f5c22db0eb1b2eec4d217"} Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.171098 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-ring-data-devices\") pod \"swift-ring-rebalance-8n9sj\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.171251 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-scripts\") pod \"swift-ring-rebalance-8n9sj\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.171281 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-combined-ca-bundle\") pod \"swift-ring-rebalance-8n9sj\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.171310 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-etc-swift\") pod \"swift-storage-0\" (UID: \"53b09c3d-ced2-431c-ab5f-2c2e09c1a686\") " pod="openstack/swift-storage-0" Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.171337 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-dispersionconf\") pod \"swift-ring-rebalance-8n9sj\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.171362 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-swiftconf\") pod \"swift-ring-rebalance-8n9sj\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.171400 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-etc-swift\") pod \"swift-ring-rebalance-8n9sj\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.171538 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42cc4\" (UniqueName: \"kubernetes.io/projected/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-kube-api-access-42cc4\") pod \"swift-ring-rebalance-8n9sj\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:10:49 crc kubenswrapper[4809]: E1124 07:10:49.171826 4809 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 07:10:49 crc kubenswrapper[4809]: E1124 07:10:49.171917 4809 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 07:10:49 crc kubenswrapper[4809]: E1124 07:10:49.172055 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-etc-swift podName:53b09c3d-ced2-431c-ab5f-2c2e09c1a686 nodeName:}" failed. No retries permitted until 2025-11-24 07:10:50.172034992 +0000 UTC m=+950.072626787 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-etc-swift") pod "swift-storage-0" (UID: "53b09c3d-ced2-431c-ab5f-2c2e09c1a686") : configmap "swift-ring-files" not found Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.172945 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-ring-data-devices\") pod \"swift-ring-rebalance-8n9sj\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.173805 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-etc-swift\") pod \"swift-ring-rebalance-8n9sj\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.174430 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-scripts\") pod \"swift-ring-rebalance-8n9sj\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.177594 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-swiftconf\") pod \"swift-ring-rebalance-8n9sj\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.178015 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-combined-ca-bundle\") pod \"swift-ring-rebalance-8n9sj\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.178532 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-dispersionconf\") pod \"swift-ring-rebalance-8n9sj\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.213807 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42cc4\" (UniqueName: \"kubernetes.io/projected/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-kube-api-access-42cc4\") pod \"swift-ring-rebalance-8n9sj\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.310710 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:10:49 crc kubenswrapper[4809]: I1124 07:10:49.752734 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-8n9sj"] Nov 24 07:10:49 crc kubenswrapper[4809]: W1124 07:10:49.764674 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ad3cf13_9872_497f_b8ee_180ec8d7bbdd.slice/crio-bf58e52536c9fc164401f109548b7753ec5432968e0c8222d8277e088433e430 WatchSource:0}: Error finding container bf58e52536c9fc164401f109548b7753ec5432968e0c8222d8277e088433e430: Status 404 returned error can't find the container with id bf58e52536c9fc164401f109548b7753ec5432968e0c8222d8277e088433e430 Nov 24 07:10:50 crc kubenswrapper[4809]: I1124 07:10:50.127482 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-jlbv5" event={"ID":"8e6016ff-f68c-423c-bef0-a015064efaaf","Type":"ContainerStarted","Data":"91115c44f7164e1aac7f11221bb5bd58295543222f37253f77e9c059cb1541ff"} Nov 24 07:10:50 crc kubenswrapper[4809]: I1124 07:10:50.127883 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-jlbv5" Nov 24 07:10:50 crc kubenswrapper[4809]: I1124 07:10:50.128858 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-8n9sj" event={"ID":"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd","Type":"ContainerStarted","Data":"bf58e52536c9fc164401f109548b7753ec5432968e0c8222d8277e088433e430"} Nov 24 07:10:50 crc kubenswrapper[4809]: I1124 07:10:50.151904 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-jlbv5" podStartSLOduration=3.151879213 podStartE2EDuration="3.151879213s" podCreationTimestamp="2025-11-24 07:10:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:10:50.146122784 +0000 UTC m=+950.046714519" watchObservedRunningTime="2025-11-24 07:10:50.151879213 +0000 UTC m=+950.052470958" Nov 24 07:10:50 crc kubenswrapper[4809]: I1124 07:10:50.186412 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-etc-swift\") pod \"swift-storage-0\" (UID: \"53b09c3d-ced2-431c-ab5f-2c2e09c1a686\") " pod="openstack/swift-storage-0" Nov 24 07:10:50 crc kubenswrapper[4809]: E1124 07:10:50.186600 4809 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 07:10:50 crc kubenswrapper[4809]: E1124 07:10:50.186645 4809 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 07:10:50 crc kubenswrapper[4809]: E1124 07:10:50.186714 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-etc-swift podName:53b09c3d-ced2-431c-ab5f-2c2e09c1a686 nodeName:}" failed. No retries permitted until 2025-11-24 07:10:52.186690349 +0000 UTC m=+952.087282074 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-etc-swift") pod "swift-storage-0" (UID: "53b09c3d-ced2-431c-ab5f-2c2e09c1a686") : configmap "swift-ring-files" not found Nov 24 07:10:50 crc kubenswrapper[4809]: I1124 07:10:50.957719 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-zhvp2"] Nov 24 07:10:50 crc kubenswrapper[4809]: I1124 07:10:50.959233 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zhvp2" Nov 24 07:10:50 crc kubenswrapper[4809]: I1124 07:10:50.963145 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-zhvp2"] Nov 24 07:10:50 crc kubenswrapper[4809]: I1124 07:10:50.978609 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-6976-account-create-cl45b"] Nov 24 07:10:50 crc kubenswrapper[4809]: I1124 07:10:50.980075 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6976-account-create-cl45b" Nov 24 07:10:50 crc kubenswrapper[4809]: I1124 07:10:50.982511 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 24 07:10:50 crc kubenswrapper[4809]: I1124 07:10:50.986529 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6976-account-create-cl45b"] Nov 24 07:10:51 crc kubenswrapper[4809]: I1124 07:10:51.101986 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2adaac7b-6a74-4d88-a1f6-91c0f76c41e1-operator-scripts\") pod \"glance-6976-account-create-cl45b\" (UID: \"2adaac7b-6a74-4d88-a1f6-91c0f76c41e1\") " pod="openstack/glance-6976-account-create-cl45b" Nov 24 07:10:51 crc kubenswrapper[4809]: I1124 07:10:51.102463 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3f63bec-2b7d-475f-a402-0f2a2e93a7f4-operator-scripts\") pod \"glance-db-create-zhvp2\" (UID: \"a3f63bec-2b7d-475f-a402-0f2a2e93a7f4\") " pod="openstack/glance-db-create-zhvp2" Nov 24 07:10:51 crc kubenswrapper[4809]: I1124 07:10:51.102582 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q5sc\" (UniqueName: \"kubernetes.io/projected/a3f63bec-2b7d-475f-a402-0f2a2e93a7f4-kube-api-access-7q5sc\") pod \"glance-db-create-zhvp2\" (UID: \"a3f63bec-2b7d-475f-a402-0f2a2e93a7f4\") " pod="openstack/glance-db-create-zhvp2" Nov 24 07:10:51 crc kubenswrapper[4809]: I1124 07:10:51.102766 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxrz4\" (UniqueName: \"kubernetes.io/projected/2adaac7b-6a74-4d88-a1f6-91c0f76c41e1-kube-api-access-dxrz4\") pod \"glance-6976-account-create-cl45b\" (UID: \"2adaac7b-6a74-4d88-a1f6-91c0f76c41e1\") " pod="openstack/glance-6976-account-create-cl45b" Nov 24 07:10:51 crc kubenswrapper[4809]: I1124 07:10:51.205083 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3f63bec-2b7d-475f-a402-0f2a2e93a7f4-operator-scripts\") pod \"glance-db-create-zhvp2\" (UID: \"a3f63bec-2b7d-475f-a402-0f2a2e93a7f4\") " pod="openstack/glance-db-create-zhvp2" Nov 24 07:10:51 crc kubenswrapper[4809]: I1124 07:10:51.205137 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q5sc\" (UniqueName: \"kubernetes.io/projected/a3f63bec-2b7d-475f-a402-0f2a2e93a7f4-kube-api-access-7q5sc\") pod \"glance-db-create-zhvp2\" (UID: \"a3f63bec-2b7d-475f-a402-0f2a2e93a7f4\") " pod="openstack/glance-db-create-zhvp2" Nov 24 07:10:51 crc kubenswrapper[4809]: I1124 07:10:51.205245 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxrz4\" (UniqueName: \"kubernetes.io/projected/2adaac7b-6a74-4d88-a1f6-91c0f76c41e1-kube-api-access-dxrz4\") pod \"glance-6976-account-create-cl45b\" (UID: \"2adaac7b-6a74-4d88-a1f6-91c0f76c41e1\") " pod="openstack/glance-6976-account-create-cl45b" Nov 24 07:10:51 crc kubenswrapper[4809]: I1124 07:10:51.205341 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2adaac7b-6a74-4d88-a1f6-91c0f76c41e1-operator-scripts\") pod \"glance-6976-account-create-cl45b\" (UID: \"2adaac7b-6a74-4d88-a1f6-91c0f76c41e1\") " pod="openstack/glance-6976-account-create-cl45b" Nov 24 07:10:51 crc kubenswrapper[4809]: I1124 07:10:51.206166 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3f63bec-2b7d-475f-a402-0f2a2e93a7f4-operator-scripts\") pod \"glance-db-create-zhvp2\" (UID: \"a3f63bec-2b7d-475f-a402-0f2a2e93a7f4\") " pod="openstack/glance-db-create-zhvp2" Nov 24 07:10:51 crc kubenswrapper[4809]: I1124 07:10:51.206729 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2adaac7b-6a74-4d88-a1f6-91c0f76c41e1-operator-scripts\") pod \"glance-6976-account-create-cl45b\" (UID: \"2adaac7b-6a74-4d88-a1f6-91c0f76c41e1\") " pod="openstack/glance-6976-account-create-cl45b" Nov 24 07:10:51 crc kubenswrapper[4809]: I1124 07:10:51.224484 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q5sc\" (UniqueName: \"kubernetes.io/projected/a3f63bec-2b7d-475f-a402-0f2a2e93a7f4-kube-api-access-7q5sc\") pod \"glance-db-create-zhvp2\" (UID: \"a3f63bec-2b7d-475f-a402-0f2a2e93a7f4\") " pod="openstack/glance-db-create-zhvp2" Nov 24 07:10:51 crc kubenswrapper[4809]: I1124 07:10:51.224531 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxrz4\" (UniqueName: \"kubernetes.io/projected/2adaac7b-6a74-4d88-a1f6-91c0f76c41e1-kube-api-access-dxrz4\") pod \"glance-6976-account-create-cl45b\" (UID: \"2adaac7b-6a74-4d88-a1f6-91c0f76c41e1\") " pod="openstack/glance-6976-account-create-cl45b" Nov 24 07:10:51 crc kubenswrapper[4809]: I1124 07:10:51.282668 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zhvp2" Nov 24 07:10:51 crc kubenswrapper[4809]: I1124 07:10:51.302634 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6976-account-create-cl45b" Nov 24 07:10:52 crc kubenswrapper[4809]: I1124 07:10:52.223653 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-etc-swift\") pod \"swift-storage-0\" (UID: \"53b09c3d-ced2-431c-ab5f-2c2e09c1a686\") " pod="openstack/swift-storage-0" Nov 24 07:10:52 crc kubenswrapper[4809]: E1124 07:10:52.224303 4809 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 07:10:52 crc kubenswrapper[4809]: E1124 07:10:52.224325 4809 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 07:10:52 crc kubenswrapper[4809]: E1124 07:10:52.224382 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-etc-swift podName:53b09c3d-ced2-431c-ab5f-2c2e09c1a686 nodeName:}" failed. No retries permitted until 2025-11-24 07:10:56.224363381 +0000 UTC m=+956.124955086 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-etc-swift") pod "swift-storage-0" (UID: "53b09c3d-ced2-431c-ab5f-2c2e09c1a686") : configmap "swift-ring-files" not found Nov 24 07:10:53 crc kubenswrapper[4809]: I1124 07:10:53.150879 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-8n9sj" event={"ID":"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd","Type":"ContainerStarted","Data":"755ed4ec84af18ee710af6dc04ebda13f3b99ed8d4c19637cc728295981baaf0"} Nov 24 07:10:53 crc kubenswrapper[4809]: I1124 07:10:53.168370 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-8n9sj" podStartSLOduration=2.058708181 podStartE2EDuration="5.16835559s" podCreationTimestamp="2025-11-24 07:10:48 +0000 UTC" firstStartedPulling="2025-11-24 07:10:49.769002967 +0000 UTC m=+949.669594672" lastFinishedPulling="2025-11-24 07:10:52.878650376 +0000 UTC m=+952.779242081" observedRunningTime="2025-11-24 07:10:53.164908241 +0000 UTC m=+953.065499956" watchObservedRunningTime="2025-11-24 07:10:53.16835559 +0000 UTC m=+953.068947295" Nov 24 07:10:53 crc kubenswrapper[4809]: I1124 07:10:53.276511 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-zhvp2"] Nov 24 07:10:53 crc kubenswrapper[4809]: W1124 07:10:53.286155 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3f63bec_2b7d_475f_a402_0f2a2e93a7f4.slice/crio-9a05995ebef2df4aaf2ffc6279b317462725e4b55bc1705de39ed4925fe63981 WatchSource:0}: Error finding container 9a05995ebef2df4aaf2ffc6279b317462725e4b55bc1705de39ed4925fe63981: Status 404 returned error can't find the container with id 9a05995ebef2df4aaf2ffc6279b317462725e4b55bc1705de39ed4925fe63981 Nov 24 07:10:53 crc kubenswrapper[4809]: W1124 07:10:53.349902 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2adaac7b_6a74_4d88_a1f6_91c0f76c41e1.slice/crio-7d5caacf287951edd94f4e5553c4665221d0149038d33fcaf0444952251cccbd WatchSource:0}: Error finding container 7d5caacf287951edd94f4e5553c4665221d0149038d33fcaf0444952251cccbd: Status 404 returned error can't find the container with id 7d5caacf287951edd94f4e5553c4665221d0149038d33fcaf0444952251cccbd Nov 24 07:10:53 crc kubenswrapper[4809]: I1124 07:10:53.359037 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6976-account-create-cl45b"] Nov 24 07:10:54 crc kubenswrapper[4809]: I1124 07:10:54.171276 4809 generic.go:334] "Generic (PLEG): container finished" podID="a3f63bec-2b7d-475f-a402-0f2a2e93a7f4" containerID="547ecbbb34ecb4ef0dc9e6286529b13444930dd157383cbf643fa05d323fc7c4" exitCode=0 Nov 24 07:10:54 crc kubenswrapper[4809]: I1124 07:10:54.171695 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zhvp2" event={"ID":"a3f63bec-2b7d-475f-a402-0f2a2e93a7f4","Type":"ContainerDied","Data":"547ecbbb34ecb4ef0dc9e6286529b13444930dd157383cbf643fa05d323fc7c4"} Nov 24 07:10:54 crc kubenswrapper[4809]: I1124 07:10:54.171741 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zhvp2" event={"ID":"a3f63bec-2b7d-475f-a402-0f2a2e93a7f4","Type":"ContainerStarted","Data":"9a05995ebef2df4aaf2ffc6279b317462725e4b55bc1705de39ed4925fe63981"} Nov 24 07:10:54 crc kubenswrapper[4809]: I1124 07:10:54.174434 4809 generic.go:334] "Generic (PLEG): container finished" podID="2adaac7b-6a74-4d88-a1f6-91c0f76c41e1" containerID="de93cbe0763eed2bf3a1bc12941fb0f8bc157e41e43e377f860e7e4554eaedc7" exitCode=0 Nov 24 07:10:54 crc kubenswrapper[4809]: I1124 07:10:54.176235 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6976-account-create-cl45b" event={"ID":"2adaac7b-6a74-4d88-a1f6-91c0f76c41e1","Type":"ContainerDied","Data":"de93cbe0763eed2bf3a1bc12941fb0f8bc157e41e43e377f860e7e4554eaedc7"} Nov 24 07:10:54 crc kubenswrapper[4809]: I1124 07:10:54.176289 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6976-account-create-cl45b" event={"ID":"2adaac7b-6a74-4d88-a1f6-91c0f76c41e1","Type":"ContainerStarted","Data":"7d5caacf287951edd94f4e5553c4665221d0149038d33fcaf0444952251cccbd"} Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.219840 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-h29zw"] Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.221534 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-h29zw" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.243684 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-h29zw"] Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.322550 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bc18-account-create-nqkws"] Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.323568 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bc18-account-create-nqkws" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.325675 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.336876 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bc18-account-create-nqkws"] Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.379717 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68b4f942-93e8-4823-91c8-761cd177776e-operator-scripts\") pod \"keystone-db-create-h29zw\" (UID: \"68b4f942-93e8-4823-91c8-761cd177776e\") " pod="openstack/keystone-db-create-h29zw" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.380092 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2865\" (UniqueName: \"kubernetes.io/projected/68b4f942-93e8-4823-91c8-761cd177776e-kube-api-access-m2865\") pod \"keystone-db-create-h29zw\" (UID: \"68b4f942-93e8-4823-91c8-761cd177776e\") " pod="openstack/keystone-db-create-h29zw" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.482223 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b-operator-scripts\") pod \"keystone-bc18-account-create-nqkws\" (UID: \"ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b\") " pod="openstack/keystone-bc18-account-create-nqkws" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.482321 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2865\" (UniqueName: \"kubernetes.io/projected/68b4f942-93e8-4823-91c8-761cd177776e-kube-api-access-m2865\") pod \"keystone-db-create-h29zw\" (UID: \"68b4f942-93e8-4823-91c8-761cd177776e\") " pod="openstack/keystone-db-create-h29zw" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.482383 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68b4f942-93e8-4823-91c8-761cd177776e-operator-scripts\") pod \"keystone-db-create-h29zw\" (UID: \"68b4f942-93e8-4823-91c8-761cd177776e\") " pod="openstack/keystone-db-create-h29zw" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.482445 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwbn9\" (UniqueName: \"kubernetes.io/projected/ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b-kube-api-access-xwbn9\") pod \"keystone-bc18-account-create-nqkws\" (UID: \"ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b\") " pod="openstack/keystone-bc18-account-create-nqkws" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.483666 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68b4f942-93e8-4823-91c8-761cd177776e-operator-scripts\") pod \"keystone-db-create-h29zw\" (UID: \"68b4f942-93e8-4823-91c8-761cd177776e\") " pod="openstack/keystone-db-create-h29zw" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.502919 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2865\" (UniqueName: \"kubernetes.io/projected/68b4f942-93e8-4823-91c8-761cd177776e-kube-api-access-m2865\") pod \"keystone-db-create-h29zw\" (UID: \"68b4f942-93e8-4823-91c8-761cd177776e\") " pod="openstack/keystone-db-create-h29zw" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.543956 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-h29zw" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.584359 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b-operator-scripts\") pod \"keystone-bc18-account-create-nqkws\" (UID: \"ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b\") " pod="openstack/keystone-bc18-account-create-nqkws" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.584518 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwbn9\" (UniqueName: \"kubernetes.io/projected/ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b-kube-api-access-xwbn9\") pod \"keystone-bc18-account-create-nqkws\" (UID: \"ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b\") " pod="openstack/keystone-bc18-account-create-nqkws" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.587189 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b-operator-scripts\") pod \"keystone-bc18-account-create-nqkws\" (UID: \"ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b\") " pod="openstack/keystone-bc18-account-create-nqkws" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.604512 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwbn9\" (UniqueName: \"kubernetes.io/projected/ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b-kube-api-access-xwbn9\") pod \"keystone-bc18-account-create-nqkws\" (UID: \"ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b\") " pod="openstack/keystone-bc18-account-create-nqkws" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.673388 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bc18-account-create-nqkws" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.697355 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zhvp2" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.703857 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-xmx6m"] Nov 24 07:10:55 crc kubenswrapper[4809]: E1124 07:10:55.704249 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3f63bec-2b7d-475f-a402-0f2a2e93a7f4" containerName="mariadb-database-create" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.704260 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3f63bec-2b7d-475f-a402-0f2a2e93a7f4" containerName="mariadb-database-create" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.704431 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3f63bec-2b7d-475f-a402-0f2a2e93a7f4" containerName="mariadb-database-create" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.704938 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-xmx6m" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.715985 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-xmx6m"] Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.718507 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6976-account-create-cl45b" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.732578 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6b09-account-create-rwt7n"] Nov 24 07:10:55 crc kubenswrapper[4809]: E1124 07:10:55.733513 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2adaac7b-6a74-4d88-a1f6-91c0f76c41e1" containerName="mariadb-account-create" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.733532 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2adaac7b-6a74-4d88-a1f6-91c0f76c41e1" containerName="mariadb-account-create" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.733768 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2adaac7b-6a74-4d88-a1f6-91c0f76c41e1" containerName="mariadb-account-create" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.739515 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6b09-account-create-rwt7n" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.740832 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.744191 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6b09-account-create-rwt7n"] Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.793500 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3f63bec-2b7d-475f-a402-0f2a2e93a7f4-operator-scripts\") pod \"a3f63bec-2b7d-475f-a402-0f2a2e93a7f4\" (UID: \"a3f63bec-2b7d-475f-a402-0f2a2e93a7f4\") " Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.793583 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7q5sc\" (UniqueName: \"kubernetes.io/projected/a3f63bec-2b7d-475f-a402-0f2a2e93a7f4-kube-api-access-7q5sc\") pod \"a3f63bec-2b7d-475f-a402-0f2a2e93a7f4\" (UID: \"a3f63bec-2b7d-475f-a402-0f2a2e93a7f4\") " Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.793948 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4a7a40c-c36f-4d21-bdfa-6035613e6719-operator-scripts\") pod \"placement-db-create-xmx6m\" (UID: \"f4a7a40c-c36f-4d21-bdfa-6035613e6719\") " pod="openstack/placement-db-create-xmx6m" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.794082 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7bmb\" (UniqueName: \"kubernetes.io/projected/f4a7a40c-c36f-4d21-bdfa-6035613e6719-kube-api-access-r7bmb\") pod \"placement-db-create-xmx6m\" (UID: \"f4a7a40c-c36f-4d21-bdfa-6035613e6719\") " pod="openstack/placement-db-create-xmx6m" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.794344 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3f63bec-2b7d-475f-a402-0f2a2e93a7f4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a3f63bec-2b7d-475f-a402-0f2a2e93a7f4" (UID: "a3f63bec-2b7d-475f-a402-0f2a2e93a7f4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.799458 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3f63bec-2b7d-475f-a402-0f2a2e93a7f4-kube-api-access-7q5sc" (OuterVolumeSpecName: "kube-api-access-7q5sc") pod "a3f63bec-2b7d-475f-a402-0f2a2e93a7f4" (UID: "a3f63bec-2b7d-475f-a402-0f2a2e93a7f4"). InnerVolumeSpecName "kube-api-access-7q5sc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.895640 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxrz4\" (UniqueName: \"kubernetes.io/projected/2adaac7b-6a74-4d88-a1f6-91c0f76c41e1-kube-api-access-dxrz4\") pod \"2adaac7b-6a74-4d88-a1f6-91c0f76c41e1\" (UID: \"2adaac7b-6a74-4d88-a1f6-91c0f76c41e1\") " Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.895690 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2adaac7b-6a74-4d88-a1f6-91c0f76c41e1-operator-scripts\") pod \"2adaac7b-6a74-4d88-a1f6-91c0f76c41e1\" (UID: \"2adaac7b-6a74-4d88-a1f6-91c0f76c41e1\") " Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.895998 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4a7a40c-c36f-4d21-bdfa-6035613e6719-operator-scripts\") pod \"placement-db-create-xmx6m\" (UID: \"f4a7a40c-c36f-4d21-bdfa-6035613e6719\") " pod="openstack/placement-db-create-xmx6m" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.896131 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27efd636-9d30-489d-aca8-f4d02581e2a8-operator-scripts\") pod \"placement-6b09-account-create-rwt7n\" (UID: \"27efd636-9d30-489d-aca8-f4d02581e2a8\") " pod="openstack/placement-6b09-account-create-rwt7n" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.896159 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7bmb\" (UniqueName: \"kubernetes.io/projected/f4a7a40c-c36f-4d21-bdfa-6035613e6719-kube-api-access-r7bmb\") pod \"placement-db-create-xmx6m\" (UID: \"f4a7a40c-c36f-4d21-bdfa-6035613e6719\") " pod="openstack/placement-db-create-xmx6m" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.896189 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2adaac7b-6a74-4d88-a1f6-91c0f76c41e1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2adaac7b-6a74-4d88-a1f6-91c0f76c41e1" (UID: "2adaac7b-6a74-4d88-a1f6-91c0f76c41e1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.896202 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpfjz\" (UniqueName: \"kubernetes.io/projected/27efd636-9d30-489d-aca8-f4d02581e2a8-kube-api-access-lpfjz\") pod \"placement-6b09-account-create-rwt7n\" (UID: \"27efd636-9d30-489d-aca8-f4d02581e2a8\") " pod="openstack/placement-6b09-account-create-rwt7n" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.896346 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3f63bec-2b7d-475f-a402-0f2a2e93a7f4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.896364 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7q5sc\" (UniqueName: \"kubernetes.io/projected/a3f63bec-2b7d-475f-a402-0f2a2e93a7f4-kube-api-access-7q5sc\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.896385 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2adaac7b-6a74-4d88-a1f6-91c0f76c41e1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.896867 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4a7a40c-c36f-4d21-bdfa-6035613e6719-operator-scripts\") pod \"placement-db-create-xmx6m\" (UID: \"f4a7a40c-c36f-4d21-bdfa-6035613e6719\") " pod="openstack/placement-db-create-xmx6m" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.899088 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2adaac7b-6a74-4d88-a1f6-91c0f76c41e1-kube-api-access-dxrz4" (OuterVolumeSpecName: "kube-api-access-dxrz4") pod "2adaac7b-6a74-4d88-a1f6-91c0f76c41e1" (UID: "2adaac7b-6a74-4d88-a1f6-91c0f76c41e1"). InnerVolumeSpecName "kube-api-access-dxrz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.925343 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7bmb\" (UniqueName: \"kubernetes.io/projected/f4a7a40c-c36f-4d21-bdfa-6035613e6719-kube-api-access-r7bmb\") pod \"placement-db-create-xmx6m\" (UID: \"f4a7a40c-c36f-4d21-bdfa-6035613e6719\") " pod="openstack/placement-db-create-xmx6m" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.998447 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpfjz\" (UniqueName: \"kubernetes.io/projected/27efd636-9d30-489d-aca8-f4d02581e2a8-kube-api-access-lpfjz\") pod \"placement-6b09-account-create-rwt7n\" (UID: \"27efd636-9d30-489d-aca8-f4d02581e2a8\") " pod="openstack/placement-6b09-account-create-rwt7n" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.998620 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27efd636-9d30-489d-aca8-f4d02581e2a8-operator-scripts\") pod \"placement-6b09-account-create-rwt7n\" (UID: \"27efd636-9d30-489d-aca8-f4d02581e2a8\") " pod="openstack/placement-6b09-account-create-rwt7n" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.998666 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxrz4\" (UniqueName: \"kubernetes.io/projected/2adaac7b-6a74-4d88-a1f6-91c0f76c41e1-kube-api-access-dxrz4\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:55 crc kubenswrapper[4809]: I1124 07:10:55.999751 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27efd636-9d30-489d-aca8-f4d02581e2a8-operator-scripts\") pod \"placement-6b09-account-create-rwt7n\" (UID: \"27efd636-9d30-489d-aca8-f4d02581e2a8\") " pod="openstack/placement-6b09-account-create-rwt7n" Nov 24 07:10:56 crc kubenswrapper[4809]: I1124 07:10:56.013747 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpfjz\" (UniqueName: \"kubernetes.io/projected/27efd636-9d30-489d-aca8-f4d02581e2a8-kube-api-access-lpfjz\") pod \"placement-6b09-account-create-rwt7n\" (UID: \"27efd636-9d30-489d-aca8-f4d02581e2a8\") " pod="openstack/placement-6b09-account-create-rwt7n" Nov 24 07:10:56 crc kubenswrapper[4809]: I1124 07:10:56.040088 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-xmx6m" Nov 24 07:10:56 crc kubenswrapper[4809]: W1124 07:10:56.046625 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68b4f942_93e8_4823_91c8_761cd177776e.slice/crio-dc18f6b820cfe36428fb0e43d2861649fd4219055aa5d0910443abd26632118e WatchSource:0}: Error finding container dc18f6b820cfe36428fb0e43d2861649fd4219055aa5d0910443abd26632118e: Status 404 returned error can't find the container with id dc18f6b820cfe36428fb0e43d2861649fd4219055aa5d0910443abd26632118e Nov 24 07:10:56 crc kubenswrapper[4809]: I1124 07:10:56.047102 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-h29zw"] Nov 24 07:10:56 crc kubenswrapper[4809]: I1124 07:10:56.064109 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6b09-account-create-rwt7n" Nov 24 07:10:56 crc kubenswrapper[4809]: I1124 07:10:56.130559 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bc18-account-create-nqkws"] Nov 24 07:10:56 crc kubenswrapper[4809]: W1124 07:10:56.188831 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef2c0d94_2e0d_4bcf_a336_f0c20fb81f2b.slice/crio-a28582de1f3043eec1217523344f1c47465221744688343d33d9e48e16f5249b WatchSource:0}: Error finding container a28582de1f3043eec1217523344f1c47465221744688343d33d9e48e16f5249b: Status 404 returned error can't find the container with id a28582de1f3043eec1217523344f1c47465221744688343d33d9e48e16f5249b Nov 24 07:10:56 crc kubenswrapper[4809]: I1124 07:10:56.198897 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6976-account-create-cl45b" event={"ID":"2adaac7b-6a74-4d88-a1f6-91c0f76c41e1","Type":"ContainerDied","Data":"7d5caacf287951edd94f4e5553c4665221d0149038d33fcaf0444952251cccbd"} Nov 24 07:10:56 crc kubenswrapper[4809]: I1124 07:10:56.198952 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d5caacf287951edd94f4e5553c4665221d0149038d33fcaf0444952251cccbd" Nov 24 07:10:56 crc kubenswrapper[4809]: I1124 07:10:56.198907 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6976-account-create-cl45b" Nov 24 07:10:56 crc kubenswrapper[4809]: I1124 07:10:56.201332 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-h29zw" event={"ID":"68b4f942-93e8-4823-91c8-761cd177776e","Type":"ContainerStarted","Data":"dc18f6b820cfe36428fb0e43d2861649fd4219055aa5d0910443abd26632118e"} Nov 24 07:10:56 crc kubenswrapper[4809]: I1124 07:10:56.209724 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zhvp2" event={"ID":"a3f63bec-2b7d-475f-a402-0f2a2e93a7f4","Type":"ContainerDied","Data":"9a05995ebef2df4aaf2ffc6279b317462725e4b55bc1705de39ed4925fe63981"} Nov 24 07:10:56 crc kubenswrapper[4809]: I1124 07:10:56.209764 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a05995ebef2df4aaf2ffc6279b317462725e4b55bc1705de39ed4925fe63981" Nov 24 07:10:56 crc kubenswrapper[4809]: I1124 07:10:56.209815 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zhvp2" Nov 24 07:10:56 crc kubenswrapper[4809]: I1124 07:10:56.302984 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-etc-swift\") pod \"swift-storage-0\" (UID: \"53b09c3d-ced2-431c-ab5f-2c2e09c1a686\") " pod="openstack/swift-storage-0" Nov 24 07:10:56 crc kubenswrapper[4809]: E1124 07:10:56.303150 4809 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 07:10:56 crc kubenswrapper[4809]: E1124 07:10:56.303180 4809 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 07:10:56 crc kubenswrapper[4809]: E1124 07:10:56.303257 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-etc-swift podName:53b09c3d-ced2-431c-ab5f-2c2e09c1a686 nodeName:}" failed. No retries permitted until 2025-11-24 07:11:04.303236466 +0000 UTC m=+964.203828181 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-etc-swift") pod "swift-storage-0" (UID: "53b09c3d-ced2-431c-ab5f-2c2e09c1a686") : configmap "swift-ring-files" not found Nov 24 07:10:56 crc kubenswrapper[4809]: I1124 07:10:56.509316 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-xmx6m"] Nov 24 07:10:56 crc kubenswrapper[4809]: W1124 07:10:56.580433 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4a7a40c_c36f_4d21_bdfa_6035613e6719.slice/crio-d34b4506c1a7a141bc4ef5d2eae1145e19f6442aa0820d35a83385914420e777 WatchSource:0}: Error finding container d34b4506c1a7a141bc4ef5d2eae1145e19f6442aa0820d35a83385914420e777: Status 404 returned error can't find the container with id d34b4506c1a7a141bc4ef5d2eae1145e19f6442aa0820d35a83385914420e777 Nov 24 07:10:56 crc kubenswrapper[4809]: I1124 07:10:56.583728 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6b09-account-create-rwt7n"] Nov 24 07:10:57 crc kubenswrapper[4809]: I1124 07:10:57.219155 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4a7a40c-c36f-4d21-bdfa-6035613e6719" containerID="a88e6255bc3813c009d8179e27ccda73e8cf0e4f9c957b49eb72735456cb2b5a" exitCode=0 Nov 24 07:10:57 crc kubenswrapper[4809]: I1124 07:10:57.219253 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-xmx6m" event={"ID":"f4a7a40c-c36f-4d21-bdfa-6035613e6719","Type":"ContainerDied","Data":"a88e6255bc3813c009d8179e27ccda73e8cf0e4f9c957b49eb72735456cb2b5a"} Nov 24 07:10:57 crc kubenswrapper[4809]: I1124 07:10:57.219280 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-xmx6m" event={"ID":"f4a7a40c-c36f-4d21-bdfa-6035613e6719","Type":"ContainerStarted","Data":"d34b4506c1a7a141bc4ef5d2eae1145e19f6442aa0820d35a83385914420e777"} Nov 24 07:10:57 crc kubenswrapper[4809]: I1124 07:10:57.222263 4809 generic.go:334] "Generic (PLEG): container finished" podID="68b4f942-93e8-4823-91c8-761cd177776e" containerID="190478af5ae637ae6c3799fd356d7d7ede679bb30cde7468370245ea96affb38" exitCode=0 Nov 24 07:10:57 crc kubenswrapper[4809]: I1124 07:10:57.222368 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-h29zw" event={"ID":"68b4f942-93e8-4823-91c8-761cd177776e","Type":"ContainerDied","Data":"190478af5ae637ae6c3799fd356d7d7ede679bb30cde7468370245ea96affb38"} Nov 24 07:10:57 crc kubenswrapper[4809]: I1124 07:10:57.225708 4809 generic.go:334] "Generic (PLEG): container finished" podID="ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b" containerID="a166853a205f39cf00e12118166298d1e418274b26182a64021afd85f5c39708" exitCode=0 Nov 24 07:10:57 crc kubenswrapper[4809]: I1124 07:10:57.225799 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bc18-account-create-nqkws" event={"ID":"ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b","Type":"ContainerDied","Data":"a166853a205f39cf00e12118166298d1e418274b26182a64021afd85f5c39708"} Nov 24 07:10:57 crc kubenswrapper[4809]: I1124 07:10:57.225833 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bc18-account-create-nqkws" event={"ID":"ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b","Type":"ContainerStarted","Data":"a28582de1f3043eec1217523344f1c47465221744688343d33d9e48e16f5249b"} Nov 24 07:10:57 crc kubenswrapper[4809]: I1124 07:10:57.228204 4809 generic.go:334] "Generic (PLEG): container finished" podID="27efd636-9d30-489d-aca8-f4d02581e2a8" containerID="9f6784d4d3e760400a2489904d937493c6e8a79e69dad97e61a457810819971b" exitCode=0 Nov 24 07:10:57 crc kubenswrapper[4809]: I1124 07:10:57.228266 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6b09-account-create-rwt7n" event={"ID":"27efd636-9d30-489d-aca8-f4d02581e2a8","Type":"ContainerDied","Data":"9f6784d4d3e760400a2489904d937493c6e8a79e69dad97e61a457810819971b"} Nov 24 07:10:57 crc kubenswrapper[4809]: I1124 07:10:57.228302 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6b09-account-create-rwt7n" event={"ID":"27efd636-9d30-489d-aca8-f4d02581e2a8","Type":"ContainerStarted","Data":"d2d10b82540bd738d3bd10af503d264be862274f4012af6dd98fc835264c8e75"} Nov 24 07:10:57 crc kubenswrapper[4809]: I1124 07:10:57.646159 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-jlbv5" Nov 24 07:10:57 crc kubenswrapper[4809]: I1124 07:10:57.719263 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-mqhjl"] Nov 24 07:10:57 crc kubenswrapper[4809]: I1124 07:10:57.719481 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" podUID="dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e" containerName="dnsmasq-dns" containerID="cri-o://e2e3af7c670851a4e17e2da182ad3af767d4ca7e2fd95e7e821fbe3d4b4a1b00" gracePeriod=10 Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.237092 4809 generic.go:334] "Generic (PLEG): container finished" podID="dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e" containerID="e2e3af7c670851a4e17e2da182ad3af767d4ca7e2fd95e7e821fbe3d4b4a1b00" exitCode=0 Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.237276 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" event={"ID":"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e","Type":"ContainerDied","Data":"e2e3af7c670851a4e17e2da182ad3af767d4ca7e2fd95e7e821fbe3d4b4a1b00"} Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.237647 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" event={"ID":"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e","Type":"ContainerDied","Data":"fa78fd6879fad9c918416963473c6782dbdc81434fcd4046b252642a31172d25"} Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.237670 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa78fd6879fad9c918416963473c6782dbdc81434fcd4046b252642a31172d25" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.315451 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.440925 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-ovsdbserver-nb\") pod \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\" (UID: \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\") " Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.441082 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-ovsdbserver-sb\") pod \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\" (UID: \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\") " Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.441159 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-config\") pod \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\" (UID: \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\") " Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.441212 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2pfm\" (UniqueName: \"kubernetes.io/projected/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-kube-api-access-m2pfm\") pod \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\" (UID: \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\") " Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.441233 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-dns-svc\") pod \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\" (UID: \"dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e\") " Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.459413 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-kube-api-access-m2pfm" (OuterVolumeSpecName: "kube-api-access-m2pfm") pod "dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e" (UID: "dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e"). InnerVolumeSpecName "kube-api-access-m2pfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.494239 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e" (UID: "dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.494898 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e" (UID: "dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.512160 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e" (UID: "dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.512977 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-config" (OuterVolumeSpecName: "config") pod "dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e" (UID: "dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.542794 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.542822 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.542831 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2pfm\" (UniqueName: \"kubernetes.io/projected/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-kube-api-access-m2pfm\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.542841 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.542852 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.547379 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6b09-account-create-rwt7n" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.644021 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27efd636-9d30-489d-aca8-f4d02581e2a8-operator-scripts\") pod \"27efd636-9d30-489d-aca8-f4d02581e2a8\" (UID: \"27efd636-9d30-489d-aca8-f4d02581e2a8\") " Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.644179 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpfjz\" (UniqueName: \"kubernetes.io/projected/27efd636-9d30-489d-aca8-f4d02581e2a8-kube-api-access-lpfjz\") pod \"27efd636-9d30-489d-aca8-f4d02581e2a8\" (UID: \"27efd636-9d30-489d-aca8-f4d02581e2a8\") " Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.645203 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27efd636-9d30-489d-aca8-f4d02581e2a8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "27efd636-9d30-489d-aca8-f4d02581e2a8" (UID: "27efd636-9d30-489d-aca8-f4d02581e2a8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.649082 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27efd636-9d30-489d-aca8-f4d02581e2a8-kube-api-access-lpfjz" (OuterVolumeSpecName: "kube-api-access-lpfjz") pod "27efd636-9d30-489d-aca8-f4d02581e2a8" (UID: "27efd636-9d30-489d-aca8-f4d02581e2a8"). InnerVolumeSpecName "kube-api-access-lpfjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.697254 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-xmx6m" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.697305 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.701770 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-h29zw" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.716748 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bc18-account-create-nqkws" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.749660 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpfjz\" (UniqueName: \"kubernetes.io/projected/27efd636-9d30-489d-aca8-f4d02581e2a8-kube-api-access-lpfjz\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.750068 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27efd636-9d30-489d-aca8-f4d02581e2a8-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.851329 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68b4f942-93e8-4823-91c8-761cd177776e-operator-scripts\") pod \"68b4f942-93e8-4823-91c8-761cd177776e\" (UID: \"68b4f942-93e8-4823-91c8-761cd177776e\") " Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.851423 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwbn9\" (UniqueName: \"kubernetes.io/projected/ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b-kube-api-access-xwbn9\") pod \"ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b\" (UID: \"ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b\") " Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.851483 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7bmb\" (UniqueName: \"kubernetes.io/projected/f4a7a40c-c36f-4d21-bdfa-6035613e6719-kube-api-access-r7bmb\") pod \"f4a7a40c-c36f-4d21-bdfa-6035613e6719\" (UID: \"f4a7a40c-c36f-4d21-bdfa-6035613e6719\") " Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.851532 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b-operator-scripts\") pod \"ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b\" (UID: \"ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b\") " Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.851559 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2865\" (UniqueName: \"kubernetes.io/projected/68b4f942-93e8-4823-91c8-761cd177776e-kube-api-access-m2865\") pod \"68b4f942-93e8-4823-91c8-761cd177776e\" (UID: \"68b4f942-93e8-4823-91c8-761cd177776e\") " Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.851579 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4a7a40c-c36f-4d21-bdfa-6035613e6719-operator-scripts\") pod \"f4a7a40c-c36f-4d21-bdfa-6035613e6719\" (UID: \"f4a7a40c-c36f-4d21-bdfa-6035613e6719\") " Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.852515 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4a7a40c-c36f-4d21-bdfa-6035613e6719-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f4a7a40c-c36f-4d21-bdfa-6035613e6719" (UID: "f4a7a40c-c36f-4d21-bdfa-6035613e6719"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.852567 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68b4f942-93e8-4823-91c8-761cd177776e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "68b4f942-93e8-4823-91c8-761cd177776e" (UID: "68b4f942-93e8-4823-91c8-761cd177776e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.852639 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b" (UID: "ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.854806 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68b4f942-93e8-4823-91c8-761cd177776e-kube-api-access-m2865" (OuterVolumeSpecName: "kube-api-access-m2865") pod "68b4f942-93e8-4823-91c8-761cd177776e" (UID: "68b4f942-93e8-4823-91c8-761cd177776e"). InnerVolumeSpecName "kube-api-access-m2865". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.855406 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b-kube-api-access-xwbn9" (OuterVolumeSpecName: "kube-api-access-xwbn9") pod "ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b" (UID: "ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b"). InnerVolumeSpecName "kube-api-access-xwbn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.856296 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4a7a40c-c36f-4d21-bdfa-6035613e6719-kube-api-access-r7bmb" (OuterVolumeSpecName: "kube-api-access-r7bmb") pod "f4a7a40c-c36f-4d21-bdfa-6035613e6719" (UID: "f4a7a40c-c36f-4d21-bdfa-6035613e6719"). InnerVolumeSpecName "kube-api-access-r7bmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.953379 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwbn9\" (UniqueName: \"kubernetes.io/projected/ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b-kube-api-access-xwbn9\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.953409 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7bmb\" (UniqueName: \"kubernetes.io/projected/f4a7a40c-c36f-4d21-bdfa-6035613e6719-kube-api-access-r7bmb\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.953421 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.953430 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2865\" (UniqueName: \"kubernetes.io/projected/68b4f942-93e8-4823-91c8-761cd177776e-kube-api-access-m2865\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.953440 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4a7a40c-c36f-4d21-bdfa-6035613e6719-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:58 crc kubenswrapper[4809]: I1124 07:10:58.953449 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68b4f942-93e8-4823-91c8-761cd177776e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:10:59 crc kubenswrapper[4809]: I1124 07:10:59.247213 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-xmx6m" event={"ID":"f4a7a40c-c36f-4d21-bdfa-6035613e6719","Type":"ContainerDied","Data":"d34b4506c1a7a141bc4ef5d2eae1145e19f6442aa0820d35a83385914420e777"} Nov 24 07:10:59 crc kubenswrapper[4809]: I1124 07:10:59.247273 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d34b4506c1a7a141bc4ef5d2eae1145e19f6442aa0820d35a83385914420e777" Nov 24 07:10:59 crc kubenswrapper[4809]: I1124 07:10:59.247244 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-xmx6m" Nov 24 07:10:59 crc kubenswrapper[4809]: I1124 07:10:59.248661 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-h29zw" event={"ID":"68b4f942-93e8-4823-91c8-761cd177776e","Type":"ContainerDied","Data":"dc18f6b820cfe36428fb0e43d2861649fd4219055aa5d0910443abd26632118e"} Nov 24 07:10:59 crc kubenswrapper[4809]: I1124 07:10:59.248705 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc18f6b820cfe36428fb0e43d2861649fd4219055aa5d0910443abd26632118e" Nov 24 07:10:59 crc kubenswrapper[4809]: I1124 07:10:59.248679 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-h29zw" Nov 24 07:10:59 crc kubenswrapper[4809]: I1124 07:10:59.250730 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bc18-account-create-nqkws" event={"ID":"ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b","Type":"ContainerDied","Data":"a28582de1f3043eec1217523344f1c47465221744688343d33d9e48e16f5249b"} Nov 24 07:10:59 crc kubenswrapper[4809]: I1124 07:10:59.250767 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a28582de1f3043eec1217523344f1c47465221744688343d33d9e48e16f5249b" Nov 24 07:10:59 crc kubenswrapper[4809]: I1124 07:10:59.250756 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bc18-account-create-nqkws" Nov 24 07:10:59 crc kubenswrapper[4809]: I1124 07:10:59.252609 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-mqhjl" Nov 24 07:10:59 crc kubenswrapper[4809]: I1124 07:10:59.252627 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6b09-account-create-rwt7n" event={"ID":"27efd636-9d30-489d-aca8-f4d02581e2a8","Type":"ContainerDied","Data":"d2d10b82540bd738d3bd10af503d264be862274f4012af6dd98fc835264c8e75"} Nov 24 07:10:59 crc kubenswrapper[4809]: I1124 07:10:59.252653 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2d10b82540bd738d3bd10af503d264be862274f4012af6dd98fc835264c8e75" Nov 24 07:10:59 crc kubenswrapper[4809]: I1124 07:10:59.252667 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6b09-account-create-rwt7n" Nov 24 07:10:59 crc kubenswrapper[4809]: I1124 07:10:59.282321 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-mqhjl"] Nov 24 07:10:59 crc kubenswrapper[4809]: I1124 07:10:59.288228 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-mqhjl"] Nov 24 07:11:00 crc kubenswrapper[4809]: I1124 07:11:00.260089 4809 generic.go:334] "Generic (PLEG): container finished" podID="9ad3cf13-9872-497f-b8ee-180ec8d7bbdd" containerID="755ed4ec84af18ee710af6dc04ebda13f3b99ed8d4c19637cc728295981baaf0" exitCode=0 Nov 24 07:11:00 crc kubenswrapper[4809]: I1124 07:11:00.260170 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-8n9sj" event={"ID":"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd","Type":"ContainerDied","Data":"755ed4ec84af18ee710af6dc04ebda13f3b99ed8d4c19637cc728295981baaf0"} Nov 24 07:11:00 crc kubenswrapper[4809]: I1124 07:11:00.914453 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e" path="/var/lib/kubelet/pods/dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e/volumes" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.136549 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-lfw7d"] Nov 24 07:11:01 crc kubenswrapper[4809]: E1124 07:11:01.136865 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e" containerName="init" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.136880 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e" containerName="init" Nov 24 07:11:01 crc kubenswrapper[4809]: E1124 07:11:01.136895 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b" containerName="mariadb-account-create" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.136902 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b" containerName="mariadb-account-create" Nov 24 07:11:01 crc kubenswrapper[4809]: E1124 07:11:01.136915 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4a7a40c-c36f-4d21-bdfa-6035613e6719" containerName="mariadb-database-create" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.136921 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4a7a40c-c36f-4d21-bdfa-6035613e6719" containerName="mariadb-database-create" Nov 24 07:11:01 crc kubenswrapper[4809]: E1124 07:11:01.136936 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27efd636-9d30-489d-aca8-f4d02581e2a8" containerName="mariadb-account-create" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.136942 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="27efd636-9d30-489d-aca8-f4d02581e2a8" containerName="mariadb-account-create" Nov 24 07:11:01 crc kubenswrapper[4809]: E1124 07:11:01.136958 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e" containerName="dnsmasq-dns" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.136981 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e" containerName="dnsmasq-dns" Nov 24 07:11:01 crc kubenswrapper[4809]: E1124 07:11:01.136993 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b4f942-93e8-4823-91c8-761cd177776e" containerName="mariadb-database-create" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.137000 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b4f942-93e8-4823-91c8-761cd177776e" containerName="mariadb-database-create" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.137150 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b4f942-93e8-4823-91c8-761cd177776e" containerName="mariadb-database-create" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.137161 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4a7a40c-c36f-4d21-bdfa-6035613e6719" containerName="mariadb-database-create" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.137173 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="27efd636-9d30-489d-aca8-f4d02581e2a8" containerName="mariadb-account-create" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.137187 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbd972c2-5bdf-4dd8-8a71-4ae6f6f8a85e" containerName="dnsmasq-dns" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.137196 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b" containerName="mariadb-account-create" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.137654 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-lfw7d" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.139690 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.140908 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-cwg4f" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.147442 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-lfw7d"] Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.291097 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtm8d\" (UniqueName: \"kubernetes.io/projected/952df16f-264d-49ca-9281-9f3b95f8273d-kube-api-access-wtm8d\") pod \"glance-db-sync-lfw7d\" (UID: \"952df16f-264d-49ca-9281-9f3b95f8273d\") " pod="openstack/glance-db-sync-lfw7d" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.291222 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/952df16f-264d-49ca-9281-9f3b95f8273d-combined-ca-bundle\") pod \"glance-db-sync-lfw7d\" (UID: \"952df16f-264d-49ca-9281-9f3b95f8273d\") " pod="openstack/glance-db-sync-lfw7d" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.291249 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/952df16f-264d-49ca-9281-9f3b95f8273d-db-sync-config-data\") pod \"glance-db-sync-lfw7d\" (UID: \"952df16f-264d-49ca-9281-9f3b95f8273d\") " pod="openstack/glance-db-sync-lfw7d" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.291299 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/952df16f-264d-49ca-9281-9f3b95f8273d-config-data\") pod \"glance-db-sync-lfw7d\" (UID: \"952df16f-264d-49ca-9281-9f3b95f8273d\") " pod="openstack/glance-db-sync-lfw7d" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.394429 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/952df16f-264d-49ca-9281-9f3b95f8273d-config-data\") pod \"glance-db-sync-lfw7d\" (UID: \"952df16f-264d-49ca-9281-9f3b95f8273d\") " pod="openstack/glance-db-sync-lfw7d" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.394548 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtm8d\" (UniqueName: \"kubernetes.io/projected/952df16f-264d-49ca-9281-9f3b95f8273d-kube-api-access-wtm8d\") pod \"glance-db-sync-lfw7d\" (UID: \"952df16f-264d-49ca-9281-9f3b95f8273d\") " pod="openstack/glance-db-sync-lfw7d" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.394625 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/952df16f-264d-49ca-9281-9f3b95f8273d-combined-ca-bundle\") pod \"glance-db-sync-lfw7d\" (UID: \"952df16f-264d-49ca-9281-9f3b95f8273d\") " pod="openstack/glance-db-sync-lfw7d" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.394657 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/952df16f-264d-49ca-9281-9f3b95f8273d-db-sync-config-data\") pod \"glance-db-sync-lfw7d\" (UID: \"952df16f-264d-49ca-9281-9f3b95f8273d\") " pod="openstack/glance-db-sync-lfw7d" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.401732 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/952df16f-264d-49ca-9281-9f3b95f8273d-db-sync-config-data\") pod \"glance-db-sync-lfw7d\" (UID: \"952df16f-264d-49ca-9281-9f3b95f8273d\") " pod="openstack/glance-db-sync-lfw7d" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.406601 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/952df16f-264d-49ca-9281-9f3b95f8273d-config-data\") pod \"glance-db-sync-lfw7d\" (UID: \"952df16f-264d-49ca-9281-9f3b95f8273d\") " pod="openstack/glance-db-sync-lfw7d" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.411202 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/952df16f-264d-49ca-9281-9f3b95f8273d-combined-ca-bundle\") pod \"glance-db-sync-lfw7d\" (UID: \"952df16f-264d-49ca-9281-9f3b95f8273d\") " pod="openstack/glance-db-sync-lfw7d" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.414431 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtm8d\" (UniqueName: \"kubernetes.io/projected/952df16f-264d-49ca-9281-9f3b95f8273d-kube-api-access-wtm8d\") pod \"glance-db-sync-lfw7d\" (UID: \"952df16f-264d-49ca-9281-9f3b95f8273d\") " pod="openstack/glance-db-sync-lfw7d" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.454750 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-lfw7d" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.601638 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.698763 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-combined-ca-bundle\") pod \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.698822 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-swiftconf\") pod \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.698885 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-etc-swift\") pod \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.698914 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-scripts\") pod \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.698949 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-dispersionconf\") pod \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.698994 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42cc4\" (UniqueName: \"kubernetes.io/projected/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-kube-api-access-42cc4\") pod \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.699031 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-ring-data-devices\") pod \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\" (UID: \"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd\") " Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.700151 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "9ad3cf13-9872-497f-b8ee-180ec8d7bbdd" (UID: "9ad3cf13-9872-497f-b8ee-180ec8d7bbdd"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.704601 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "9ad3cf13-9872-497f-b8ee-180ec8d7bbdd" (UID: "9ad3cf13-9872-497f-b8ee-180ec8d7bbdd"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.705919 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-kube-api-access-42cc4" (OuterVolumeSpecName: "kube-api-access-42cc4") pod "9ad3cf13-9872-497f-b8ee-180ec8d7bbdd" (UID: "9ad3cf13-9872-497f-b8ee-180ec8d7bbdd"). InnerVolumeSpecName "kube-api-access-42cc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.706344 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "9ad3cf13-9872-497f-b8ee-180ec8d7bbdd" (UID: "9ad3cf13-9872-497f-b8ee-180ec8d7bbdd"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.717782 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-scripts" (OuterVolumeSpecName: "scripts") pod "9ad3cf13-9872-497f-b8ee-180ec8d7bbdd" (UID: "9ad3cf13-9872-497f-b8ee-180ec8d7bbdd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.720409 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "9ad3cf13-9872-497f-b8ee-180ec8d7bbdd" (UID: "9ad3cf13-9872-497f-b8ee-180ec8d7bbdd"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.721698 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9ad3cf13-9872-497f-b8ee-180ec8d7bbdd" (UID: "9ad3cf13-9872-497f-b8ee-180ec8d7bbdd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.800869 4809 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.801187 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42cc4\" (UniqueName: \"kubernetes.io/projected/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-kube-api-access-42cc4\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.801250 4809 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.801312 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.801367 4809 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.801427 4809 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.801487 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ad3cf13-9872-497f-b8ee-180ec8d7bbdd-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:01 crc kubenswrapper[4809]: I1124 07:11:01.979305 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-lfw7d"] Nov 24 07:11:02 crc kubenswrapper[4809]: I1124 07:11:02.272889 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-lfw7d" event={"ID":"952df16f-264d-49ca-9281-9f3b95f8273d","Type":"ContainerStarted","Data":"31ff16a55d2882fb1dbce8d1349259cbe85d92224b284156c87db82aad4d2182"} Nov 24 07:11:02 crc kubenswrapper[4809]: I1124 07:11:02.274470 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-8n9sj" event={"ID":"9ad3cf13-9872-497f-b8ee-180ec8d7bbdd","Type":"ContainerDied","Data":"bf58e52536c9fc164401f109548b7753ec5432968e0c8222d8277e088433e430"} Nov 24 07:11:02 crc kubenswrapper[4809]: I1124 07:11:02.274495 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf58e52536c9fc164401f109548b7753ec5432968e0c8222d8277e088433e430" Nov 24 07:11:02 crc kubenswrapper[4809]: I1124 07:11:02.274558 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-8n9sj" Nov 24 07:11:04 crc kubenswrapper[4809]: I1124 07:11:04.339245 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-etc-swift\") pod \"swift-storage-0\" (UID: \"53b09c3d-ced2-431c-ab5f-2c2e09c1a686\") " pod="openstack/swift-storage-0" Nov 24 07:11:04 crc kubenswrapper[4809]: I1124 07:11:04.349844 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/53b09c3d-ced2-431c-ab5f-2c2e09c1a686-etc-swift\") pod \"swift-storage-0\" (UID: \"53b09c3d-ced2-431c-ab5f-2c2e09c1a686\") " pod="openstack/swift-storage-0" Nov 24 07:11:04 crc kubenswrapper[4809]: I1124 07:11:04.378813 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 07:11:04 crc kubenswrapper[4809]: I1124 07:11:04.924137 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 24 07:11:05 crc kubenswrapper[4809]: I1124 07:11:05.302626 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"53b09c3d-ced2-431c-ab5f-2c2e09c1a686","Type":"ContainerStarted","Data":"f9879b81854e67510471f877509364b1cfcf63a864d19df15eac4bd3c9e199ab"} Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.219324 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-vnpwc" podUID="c14ae567-41ab-455a-8756-04daf7c81677" containerName="ovn-controller" probeResult="failure" output=< Nov 24 07:11:06 crc kubenswrapper[4809]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 07:11:06 crc kubenswrapper[4809]: > Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.251109 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.295325 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-drz48" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.313229 4809 generic.go:334] "Generic (PLEG): container finished" podID="f91f6ab4-0c7c-4002-aad9-dcd408c9000e" containerID="ae51b70d41f1cc092e89e2f506b0d859bd045728a8e37fb6553593c1785b0da8" exitCode=0 Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.313318 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f91f6ab4-0c7c-4002-aad9-dcd408c9000e","Type":"ContainerDied","Data":"ae51b70d41f1cc092e89e2f506b0d859bd045728a8e37fb6553593c1785b0da8"} Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.524134 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-vnpwc-config-tddvf"] Nov 24 07:11:06 crc kubenswrapper[4809]: E1124 07:11:06.524551 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ad3cf13-9872-497f-b8ee-180ec8d7bbdd" containerName="swift-ring-rebalance" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.524567 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ad3cf13-9872-497f-b8ee-180ec8d7bbdd" containerName="swift-ring-rebalance" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.524734 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ad3cf13-9872-497f-b8ee-180ec8d7bbdd" containerName="swift-ring-rebalance" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.525348 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vnpwc-config-tddvf" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.528065 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.532771 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vnpwc-config-tddvf"] Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.573329 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4b86aee6-8e03-47dc-bf45-34a59b964dd1-var-log-ovn\") pod \"ovn-controller-vnpwc-config-tddvf\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " pod="openstack/ovn-controller-vnpwc-config-tddvf" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.573395 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4b86aee6-8e03-47dc-bf45-34a59b964dd1-additional-scripts\") pod \"ovn-controller-vnpwc-config-tddvf\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " pod="openstack/ovn-controller-vnpwc-config-tddvf" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.573430 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b86aee6-8e03-47dc-bf45-34a59b964dd1-scripts\") pod \"ovn-controller-vnpwc-config-tddvf\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " pod="openstack/ovn-controller-vnpwc-config-tddvf" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.573457 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4b86aee6-8e03-47dc-bf45-34a59b964dd1-var-run\") pod \"ovn-controller-vnpwc-config-tddvf\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " pod="openstack/ovn-controller-vnpwc-config-tddvf" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.573526 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94v65\" (UniqueName: \"kubernetes.io/projected/4b86aee6-8e03-47dc-bf45-34a59b964dd1-kube-api-access-94v65\") pod \"ovn-controller-vnpwc-config-tddvf\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " pod="openstack/ovn-controller-vnpwc-config-tddvf" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.573561 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4b86aee6-8e03-47dc-bf45-34a59b964dd1-var-run-ovn\") pod \"ovn-controller-vnpwc-config-tddvf\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " pod="openstack/ovn-controller-vnpwc-config-tddvf" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.674812 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94v65\" (UniqueName: \"kubernetes.io/projected/4b86aee6-8e03-47dc-bf45-34a59b964dd1-kube-api-access-94v65\") pod \"ovn-controller-vnpwc-config-tddvf\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " pod="openstack/ovn-controller-vnpwc-config-tddvf" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.675197 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4b86aee6-8e03-47dc-bf45-34a59b964dd1-var-run-ovn\") pod \"ovn-controller-vnpwc-config-tddvf\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " pod="openstack/ovn-controller-vnpwc-config-tddvf" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.675298 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4b86aee6-8e03-47dc-bf45-34a59b964dd1-var-log-ovn\") pod \"ovn-controller-vnpwc-config-tddvf\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " pod="openstack/ovn-controller-vnpwc-config-tddvf" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.675338 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4b86aee6-8e03-47dc-bf45-34a59b964dd1-additional-scripts\") pod \"ovn-controller-vnpwc-config-tddvf\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " pod="openstack/ovn-controller-vnpwc-config-tddvf" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.675364 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b86aee6-8e03-47dc-bf45-34a59b964dd1-scripts\") pod \"ovn-controller-vnpwc-config-tddvf\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " pod="openstack/ovn-controller-vnpwc-config-tddvf" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.675391 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4b86aee6-8e03-47dc-bf45-34a59b964dd1-var-run\") pod \"ovn-controller-vnpwc-config-tddvf\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " pod="openstack/ovn-controller-vnpwc-config-tddvf" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.675756 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4b86aee6-8e03-47dc-bf45-34a59b964dd1-var-run\") pod \"ovn-controller-vnpwc-config-tddvf\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " pod="openstack/ovn-controller-vnpwc-config-tddvf" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.675830 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4b86aee6-8e03-47dc-bf45-34a59b964dd1-var-log-ovn\") pod \"ovn-controller-vnpwc-config-tddvf\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " pod="openstack/ovn-controller-vnpwc-config-tddvf" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.675880 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4b86aee6-8e03-47dc-bf45-34a59b964dd1-var-run-ovn\") pod \"ovn-controller-vnpwc-config-tddvf\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " pod="openstack/ovn-controller-vnpwc-config-tddvf" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.676669 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4b86aee6-8e03-47dc-bf45-34a59b964dd1-additional-scripts\") pod \"ovn-controller-vnpwc-config-tddvf\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " pod="openstack/ovn-controller-vnpwc-config-tddvf" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.680761 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b86aee6-8e03-47dc-bf45-34a59b964dd1-scripts\") pod \"ovn-controller-vnpwc-config-tddvf\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " pod="openstack/ovn-controller-vnpwc-config-tddvf" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.702524 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94v65\" (UniqueName: \"kubernetes.io/projected/4b86aee6-8e03-47dc-bf45-34a59b964dd1-kube-api-access-94v65\") pod \"ovn-controller-vnpwc-config-tddvf\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " pod="openstack/ovn-controller-vnpwc-config-tddvf" Nov 24 07:11:06 crc kubenswrapper[4809]: I1124 07:11:06.854694 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vnpwc-config-tddvf" Nov 24 07:11:07 crc kubenswrapper[4809]: I1124 07:11:07.322992 4809 generic.go:334] "Generic (PLEG): container finished" podID="7d0a537c-b1f0-4908-a604-424ea6a58438" containerID="3e28454c296359bd8a170cd1fd9c49436203c71b0aad0ea763826736402a4456" exitCode=0 Nov 24 07:11:07 crc kubenswrapper[4809]: I1124 07:11:07.323035 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7d0a537c-b1f0-4908-a604-424ea6a58438","Type":"ContainerDied","Data":"3e28454c296359bd8a170cd1fd9c49436203c71b0aad0ea763826736402a4456"} Nov 24 07:11:11 crc kubenswrapper[4809]: I1124 07:11:11.215462 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-vnpwc" podUID="c14ae567-41ab-455a-8756-04daf7c81677" containerName="ovn-controller" probeResult="failure" output=< Nov 24 07:11:11 crc kubenswrapper[4809]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 07:11:11 crc kubenswrapper[4809]: > Nov 24 07:11:12 crc kubenswrapper[4809]: I1124 07:11:12.908693 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vnpwc-config-tddvf"] Nov 24 07:11:12 crc kubenswrapper[4809]: W1124 07:11:12.920449 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b86aee6_8e03_47dc_bf45_34a59b964dd1.slice/crio-e0425644c12b77b4c710927758d7d3b901dc8896cef8466a37ba5ba53739aa24 WatchSource:0}: Error finding container e0425644c12b77b4c710927758d7d3b901dc8896cef8466a37ba5ba53739aa24: Status 404 returned error can't find the container with id e0425644c12b77b4c710927758d7d3b901dc8896cef8466a37ba5ba53739aa24 Nov 24 07:11:13 crc kubenswrapper[4809]: I1124 07:11:13.367763 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f91f6ab4-0c7c-4002-aad9-dcd408c9000e","Type":"ContainerStarted","Data":"d6cfe70d34456cc68d198ac9f1cb919fdebac98b1adbc19e4e8accb2affa5c81"} Nov 24 07:11:13 crc kubenswrapper[4809]: I1124 07:11:13.368286 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:11:13 crc kubenswrapper[4809]: I1124 07:11:13.370949 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"53b09c3d-ced2-431c-ab5f-2c2e09c1a686","Type":"ContainerStarted","Data":"458fb7c20a800c92c7947c1bd3ac162dc7ff2ce6ea6b086018c1b7e253d067ca"} Nov 24 07:11:13 crc kubenswrapper[4809]: I1124 07:11:13.371017 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"53b09c3d-ced2-431c-ab5f-2c2e09c1a686","Type":"ContainerStarted","Data":"91c06a8e3b2dd575a89a4f6841c1ce471e2999f1b3b99397d3d59af3a59f905f"} Nov 24 07:11:13 crc kubenswrapper[4809]: I1124 07:11:13.371030 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"53b09c3d-ced2-431c-ab5f-2c2e09c1a686","Type":"ContainerStarted","Data":"ec07bceb6215c91a525fa8e650290804c5b15b354ded6ecb6113685a64ad5f9e"} Nov 24 07:11:13 crc kubenswrapper[4809]: I1124 07:11:13.371039 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"53b09c3d-ced2-431c-ab5f-2c2e09c1a686","Type":"ContainerStarted","Data":"17661204655a701c17d12e482a3e9012970b2fe4260a4272849b1b221a7252e9"} Nov 24 07:11:13 crc kubenswrapper[4809]: I1124 07:11:13.372718 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7d0a537c-b1f0-4908-a604-424ea6a58438","Type":"ContainerStarted","Data":"20a05ddef049ee32e97c296b2435378bf343ba8c54800ac573d5e1caa2bfa200"} Nov 24 07:11:13 crc kubenswrapper[4809]: I1124 07:11:13.373020 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 07:11:13 crc kubenswrapper[4809]: I1124 07:11:13.374932 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vnpwc-config-tddvf" event={"ID":"4b86aee6-8e03-47dc-bf45-34a59b964dd1","Type":"ContainerStarted","Data":"9847ef762be1e8b052fdd7e6a83f18c7bc41cb4c56be305015af58ce1c31776e"} Nov 24 07:11:13 crc kubenswrapper[4809]: I1124 07:11:13.375011 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vnpwc-config-tddvf" event={"ID":"4b86aee6-8e03-47dc-bf45-34a59b964dd1","Type":"ContainerStarted","Data":"e0425644c12b77b4c710927758d7d3b901dc8896cef8466a37ba5ba53739aa24"} Nov 24 07:11:13 crc kubenswrapper[4809]: I1124 07:11:13.376561 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-lfw7d" event={"ID":"952df16f-264d-49ca-9281-9f3b95f8273d","Type":"ContainerStarted","Data":"9e6f28cb3c7b8bc7f9374a06b3bc896f2892963f6b563dd75c5d836ebfb60db8"} Nov 24 07:11:13 crc kubenswrapper[4809]: I1124 07:11:13.397332 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=54.585295281 podStartE2EDuration="1m2.39731337s" podCreationTimestamp="2025-11-24 07:10:11 +0000 UTC" firstStartedPulling="2025-11-24 07:10:21.931151313 +0000 UTC m=+921.831743018" lastFinishedPulling="2025-11-24 07:10:29.743169412 +0000 UTC m=+929.643761107" observedRunningTime="2025-11-24 07:11:13.393623152 +0000 UTC m=+973.294214857" watchObservedRunningTime="2025-11-24 07:11:13.39731337 +0000 UTC m=+973.297905075" Nov 24 07:11:13 crc kubenswrapper[4809]: I1124 07:11:13.419389 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=54.508205068 podStartE2EDuration="1m3.419367194s" podCreationTimestamp="2025-11-24 07:10:10 +0000 UTC" firstStartedPulling="2025-11-24 07:10:22.871188289 +0000 UTC m=+922.771779994" lastFinishedPulling="2025-11-24 07:10:31.782350415 +0000 UTC m=+931.682942120" observedRunningTime="2025-11-24 07:11:13.415504252 +0000 UTC m=+973.316095977" watchObservedRunningTime="2025-11-24 07:11:13.419367194 +0000 UTC m=+973.319958919" Nov 24 07:11:13 crc kubenswrapper[4809]: I1124 07:11:13.433364 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-lfw7d" podStartSLOduration=1.888357107 podStartE2EDuration="12.433342265s" podCreationTimestamp="2025-11-24 07:11:01 +0000 UTC" firstStartedPulling="2025-11-24 07:11:01.983089517 +0000 UTC m=+961.883681222" lastFinishedPulling="2025-11-24 07:11:12.528074675 +0000 UTC m=+972.428666380" observedRunningTime="2025-11-24 07:11:13.427759007 +0000 UTC m=+973.328350712" watchObservedRunningTime="2025-11-24 07:11:13.433342265 +0000 UTC m=+973.333933970" Nov 24 07:11:13 crc kubenswrapper[4809]: I1124 07:11:13.442867 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-vnpwc-config-tddvf" podStartSLOduration=7.442852888 podStartE2EDuration="7.442852888s" podCreationTimestamp="2025-11-24 07:11:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:11:13.440255669 +0000 UTC m=+973.340847374" watchObservedRunningTime="2025-11-24 07:11:13.442852888 +0000 UTC m=+973.343444593" Nov 24 07:11:14 crc kubenswrapper[4809]: I1124 07:11:14.389127 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"53b09c3d-ced2-431c-ab5f-2c2e09c1a686","Type":"ContainerStarted","Data":"e48c70d10adb479b4d8c5285068cc75fd54d4bbed4f4d69b6f0ec8c79b9c9001"} Nov 24 07:11:14 crc kubenswrapper[4809]: I1124 07:11:14.391310 4809 generic.go:334] "Generic (PLEG): container finished" podID="4b86aee6-8e03-47dc-bf45-34a59b964dd1" containerID="9847ef762be1e8b052fdd7e6a83f18c7bc41cb4c56be305015af58ce1c31776e" exitCode=0 Nov 24 07:11:14 crc kubenswrapper[4809]: I1124 07:11:14.391440 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vnpwc-config-tddvf" event={"ID":"4b86aee6-8e03-47dc-bf45-34a59b964dd1","Type":"ContainerDied","Data":"9847ef762be1e8b052fdd7e6a83f18c7bc41cb4c56be305015af58ce1c31776e"} Nov 24 07:11:15 crc kubenswrapper[4809]: I1124 07:11:15.406081 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"53b09c3d-ced2-431c-ab5f-2c2e09c1a686","Type":"ContainerStarted","Data":"014e2e0ba5a149e7c4c7d053c19d54dc141721d6e175d47c5456027143888943"} Nov 24 07:11:15 crc kubenswrapper[4809]: I1124 07:11:15.406428 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"53b09c3d-ced2-431c-ab5f-2c2e09c1a686","Type":"ContainerStarted","Data":"fde70777582e2313342c7582e916e9ccd6d63153bfb6ea400803c095124b5a46"} Nov 24 07:11:15 crc kubenswrapper[4809]: I1124 07:11:15.406442 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"53b09c3d-ced2-431c-ab5f-2c2e09c1a686","Type":"ContainerStarted","Data":"36585ad93c0f80064b93feae2c5cd1f94a911765e6b09feffcd1281be3e151cb"} Nov 24 07:11:15 crc kubenswrapper[4809]: I1124 07:11:15.839020 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vnpwc-config-tddvf" Nov 24 07:11:15 crc kubenswrapper[4809]: I1124 07:11:15.929757 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4b86aee6-8e03-47dc-bf45-34a59b964dd1-var-run\") pod \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " Nov 24 07:11:15 crc kubenswrapper[4809]: I1124 07:11:15.929855 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4b86aee6-8e03-47dc-bf45-34a59b964dd1-var-log-ovn\") pod \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " Nov 24 07:11:15 crc kubenswrapper[4809]: I1124 07:11:15.929853 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4b86aee6-8e03-47dc-bf45-34a59b964dd1-var-run" (OuterVolumeSpecName: "var-run") pod "4b86aee6-8e03-47dc-bf45-34a59b964dd1" (UID: "4b86aee6-8e03-47dc-bf45-34a59b964dd1"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:11:15 crc kubenswrapper[4809]: I1124 07:11:15.929898 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4b86aee6-8e03-47dc-bf45-34a59b964dd1-additional-scripts\") pod \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " Nov 24 07:11:15 crc kubenswrapper[4809]: I1124 07:11:15.930059 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b86aee6-8e03-47dc-bf45-34a59b964dd1-scripts\") pod \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " Nov 24 07:11:15 crc kubenswrapper[4809]: I1124 07:11:15.930161 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4b86aee6-8e03-47dc-bf45-34a59b964dd1-var-run-ovn\") pod \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " Nov 24 07:11:15 crc kubenswrapper[4809]: I1124 07:11:15.930238 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94v65\" (UniqueName: \"kubernetes.io/projected/4b86aee6-8e03-47dc-bf45-34a59b964dd1-kube-api-access-94v65\") pod \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\" (UID: \"4b86aee6-8e03-47dc-bf45-34a59b964dd1\") " Nov 24 07:11:15 crc kubenswrapper[4809]: I1124 07:11:15.930652 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b86aee6-8e03-47dc-bf45-34a59b964dd1-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "4b86aee6-8e03-47dc-bf45-34a59b964dd1" (UID: "4b86aee6-8e03-47dc-bf45-34a59b964dd1"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:15 crc kubenswrapper[4809]: I1124 07:11:15.930859 4809 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4b86aee6-8e03-47dc-bf45-34a59b964dd1-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:15 crc kubenswrapper[4809]: I1124 07:11:15.930876 4809 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4b86aee6-8e03-47dc-bf45-34a59b964dd1-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:15 crc kubenswrapper[4809]: I1124 07:11:15.931682 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4b86aee6-8e03-47dc-bf45-34a59b964dd1-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "4b86aee6-8e03-47dc-bf45-34a59b964dd1" (UID: "4b86aee6-8e03-47dc-bf45-34a59b964dd1"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:11:15 crc kubenswrapper[4809]: I1124 07:11:15.931729 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4b86aee6-8e03-47dc-bf45-34a59b964dd1-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "4b86aee6-8e03-47dc-bf45-34a59b964dd1" (UID: "4b86aee6-8e03-47dc-bf45-34a59b964dd1"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:11:15 crc kubenswrapper[4809]: I1124 07:11:15.932517 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b86aee6-8e03-47dc-bf45-34a59b964dd1-scripts" (OuterVolumeSpecName: "scripts") pod "4b86aee6-8e03-47dc-bf45-34a59b964dd1" (UID: "4b86aee6-8e03-47dc-bf45-34a59b964dd1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:15 crc kubenswrapper[4809]: I1124 07:11:15.946233 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b86aee6-8e03-47dc-bf45-34a59b964dd1-kube-api-access-94v65" (OuterVolumeSpecName: "kube-api-access-94v65") pod "4b86aee6-8e03-47dc-bf45-34a59b964dd1" (UID: "4b86aee6-8e03-47dc-bf45-34a59b964dd1"). InnerVolumeSpecName "kube-api-access-94v65". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.001190 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-vnpwc-config-tddvf"] Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.009829 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-vnpwc-config-tddvf"] Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.038009 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94v65\" (UniqueName: \"kubernetes.io/projected/4b86aee6-8e03-47dc-bf45-34a59b964dd1-kube-api-access-94v65\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.038214 4809 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4b86aee6-8e03-47dc-bf45-34a59b964dd1-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.038306 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b86aee6-8e03-47dc-bf45-34a59b964dd1-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.038379 4809 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4b86aee6-8e03-47dc-bf45-34a59b964dd1-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.135395 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-vnpwc-config-4nptz"] Nov 24 07:11:16 crc kubenswrapper[4809]: E1124 07:11:16.135725 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b86aee6-8e03-47dc-bf45-34a59b964dd1" containerName="ovn-config" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.135742 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b86aee6-8e03-47dc-bf45-34a59b964dd1" containerName="ovn-config" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.135956 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b86aee6-8e03-47dc-bf45-34a59b964dd1" containerName="ovn-config" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.136495 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vnpwc-config-4nptz" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.149587 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vnpwc-config-4nptz"] Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.224210 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-vnpwc" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.242490 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/167875f4-19ea-454e-8629-776ba700ac28-var-run-ovn\") pod \"ovn-controller-vnpwc-config-4nptz\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " pod="openstack/ovn-controller-vnpwc-config-4nptz" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.242807 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/167875f4-19ea-454e-8629-776ba700ac28-scripts\") pod \"ovn-controller-vnpwc-config-4nptz\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " pod="openstack/ovn-controller-vnpwc-config-4nptz" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.242829 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/167875f4-19ea-454e-8629-776ba700ac28-additional-scripts\") pod \"ovn-controller-vnpwc-config-4nptz\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " pod="openstack/ovn-controller-vnpwc-config-4nptz" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.242862 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/167875f4-19ea-454e-8629-776ba700ac28-var-log-ovn\") pod \"ovn-controller-vnpwc-config-4nptz\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " pod="openstack/ovn-controller-vnpwc-config-4nptz" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.242885 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkpv6\" (UniqueName: \"kubernetes.io/projected/167875f4-19ea-454e-8629-776ba700ac28-kube-api-access-fkpv6\") pod \"ovn-controller-vnpwc-config-4nptz\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " pod="openstack/ovn-controller-vnpwc-config-4nptz" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.243000 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/167875f4-19ea-454e-8629-776ba700ac28-var-run\") pod \"ovn-controller-vnpwc-config-4nptz\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " pod="openstack/ovn-controller-vnpwc-config-4nptz" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.344342 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/167875f4-19ea-454e-8629-776ba700ac28-var-run\") pod \"ovn-controller-vnpwc-config-4nptz\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " pod="openstack/ovn-controller-vnpwc-config-4nptz" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.344422 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/167875f4-19ea-454e-8629-776ba700ac28-var-run-ovn\") pod \"ovn-controller-vnpwc-config-4nptz\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " pod="openstack/ovn-controller-vnpwc-config-4nptz" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.344670 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/167875f4-19ea-454e-8629-776ba700ac28-var-run\") pod \"ovn-controller-vnpwc-config-4nptz\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " pod="openstack/ovn-controller-vnpwc-config-4nptz" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.344723 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/167875f4-19ea-454e-8629-776ba700ac28-scripts\") pod \"ovn-controller-vnpwc-config-4nptz\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " pod="openstack/ovn-controller-vnpwc-config-4nptz" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.344742 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/167875f4-19ea-454e-8629-776ba700ac28-additional-scripts\") pod \"ovn-controller-vnpwc-config-4nptz\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " pod="openstack/ovn-controller-vnpwc-config-4nptz" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.344756 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/167875f4-19ea-454e-8629-776ba700ac28-var-run-ovn\") pod \"ovn-controller-vnpwc-config-4nptz\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " pod="openstack/ovn-controller-vnpwc-config-4nptz" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.345387 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/167875f4-19ea-454e-8629-776ba700ac28-additional-scripts\") pod \"ovn-controller-vnpwc-config-4nptz\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " pod="openstack/ovn-controller-vnpwc-config-4nptz" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.346678 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/167875f4-19ea-454e-8629-776ba700ac28-scripts\") pod \"ovn-controller-vnpwc-config-4nptz\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " pod="openstack/ovn-controller-vnpwc-config-4nptz" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.347099 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/167875f4-19ea-454e-8629-776ba700ac28-var-log-ovn\") pod \"ovn-controller-vnpwc-config-4nptz\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " pod="openstack/ovn-controller-vnpwc-config-4nptz" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.347285 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/167875f4-19ea-454e-8629-776ba700ac28-var-log-ovn\") pod \"ovn-controller-vnpwc-config-4nptz\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " pod="openstack/ovn-controller-vnpwc-config-4nptz" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.347341 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkpv6\" (UniqueName: \"kubernetes.io/projected/167875f4-19ea-454e-8629-776ba700ac28-kube-api-access-fkpv6\") pod \"ovn-controller-vnpwc-config-4nptz\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " pod="openstack/ovn-controller-vnpwc-config-4nptz" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.366377 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkpv6\" (UniqueName: \"kubernetes.io/projected/167875f4-19ea-454e-8629-776ba700ac28-kube-api-access-fkpv6\") pod \"ovn-controller-vnpwc-config-4nptz\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " pod="openstack/ovn-controller-vnpwc-config-4nptz" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.413231 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0425644c12b77b4c710927758d7d3b901dc8896cef8466a37ba5ba53739aa24" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.413288 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vnpwc-config-tddvf" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.422709 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"53b09c3d-ced2-431c-ab5f-2c2e09c1a686","Type":"ContainerStarted","Data":"3a07e60f19042db4870269c3c9b46ed52e7b328eb8f4b541cfb9580231973bda"} Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.422762 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"53b09c3d-ced2-431c-ab5f-2c2e09c1a686","Type":"ContainerStarted","Data":"99add0fc68e03d26215de491c7cf8e2d402fa66830997b66adc9a70ab429ae00"} Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.422771 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"53b09c3d-ced2-431c-ab5f-2c2e09c1a686","Type":"ContainerStarted","Data":"0b38fe56f2bb8de885a65770b8dcc236dd439c05ff15178956a2d89d603dca95"} Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.491182 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vnpwc-config-4nptz" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.913314 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b86aee6-8e03-47dc-bf45-34a59b964dd1" path="/var/lib/kubelet/pods/4b86aee6-8e03-47dc-bf45-34a59b964dd1/volumes" Nov 24 07:11:16 crc kubenswrapper[4809]: I1124 07:11:16.959722 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vnpwc-config-4nptz"] Nov 24 07:11:16 crc kubenswrapper[4809]: W1124 07:11:16.968957 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod167875f4_19ea_454e_8629_776ba700ac28.slice/crio-3723685157d353ae927f542c0fe035efeb25a4dbc28a00bf1bec1397700f0f70 WatchSource:0}: Error finding container 3723685157d353ae927f542c0fe035efeb25a4dbc28a00bf1bec1397700f0f70: Status 404 returned error can't find the container with id 3723685157d353ae927f542c0fe035efeb25a4dbc28a00bf1bec1397700f0f70 Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.435959 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"53b09c3d-ced2-431c-ab5f-2c2e09c1a686","Type":"ContainerStarted","Data":"fcefb5b906b243d0d643dc0b1532def1ce71730bfc6ddfe415d6c55a8f3329e8"} Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.436277 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"53b09c3d-ced2-431c-ab5f-2c2e09c1a686","Type":"ContainerStarted","Data":"620897669ebc9c547189e4f28731bd5ef64447a8e559e5ff3ce7ce1c86422c3b"} Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.436288 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"53b09c3d-ced2-431c-ab5f-2c2e09c1a686","Type":"ContainerStarted","Data":"cdfb2d23b8eb0e4cfd861e6b2543b97c723976f13fbd70b0e5e85a4eaa8bb7cb"} Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.436296 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"53b09c3d-ced2-431c-ab5f-2c2e09c1a686","Type":"ContainerStarted","Data":"d8f9ed05c72e3549b6a25c34e3bfdbac1900c3fd5bb23ab50e66ae6fceceeb79"} Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.439707 4809 generic.go:334] "Generic (PLEG): container finished" podID="167875f4-19ea-454e-8629-776ba700ac28" containerID="4436bea8e3c221175588f21ce0a5d2e2af759b714bd7d68636b3cf70e819c4c4" exitCode=0 Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.439754 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vnpwc-config-4nptz" event={"ID":"167875f4-19ea-454e-8629-776ba700ac28","Type":"ContainerDied","Data":"4436bea8e3c221175588f21ce0a5d2e2af759b714bd7d68636b3cf70e819c4c4"} Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.439780 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vnpwc-config-4nptz" event={"ID":"167875f4-19ea-454e-8629-776ba700ac28","Type":"ContainerStarted","Data":"3723685157d353ae927f542c0fe035efeb25a4dbc28a00bf1bec1397700f0f70"} Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.467323 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=19.591955382 podStartE2EDuration="30.467297143s" podCreationTimestamp="2025-11-24 07:10:47 +0000 UTC" firstStartedPulling="2025-11-24 07:11:04.940750745 +0000 UTC m=+964.841342440" lastFinishedPulling="2025-11-24 07:11:15.816092496 +0000 UTC m=+975.716684201" observedRunningTime="2025-11-24 07:11:17.4638014 +0000 UTC m=+977.364393105" watchObservedRunningTime="2025-11-24 07:11:17.467297143 +0000 UTC m=+977.367888848" Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.739118 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-qq7dz"] Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.740719 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.745160 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.769308 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-qq7dz"] Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.773540 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-dns-svc\") pod \"dnsmasq-dns-764c5664d7-qq7dz\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.773643 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-qq7dz\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.773689 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-config\") pod \"dnsmasq-dns-764c5664d7-qq7dz\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.773722 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4sjs\" (UniqueName: \"kubernetes.io/projected/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-kube-api-access-q4sjs\") pod \"dnsmasq-dns-764c5664d7-qq7dz\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.773786 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-qq7dz\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.773953 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-qq7dz\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.876012 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-qq7dz\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.876090 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-config\") pod \"dnsmasq-dns-764c5664d7-qq7dz\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.876124 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4sjs\" (UniqueName: \"kubernetes.io/projected/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-kube-api-access-q4sjs\") pod \"dnsmasq-dns-764c5664d7-qq7dz\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.876196 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-qq7dz\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.876230 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-qq7dz\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.876302 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-dns-svc\") pod \"dnsmasq-dns-764c5664d7-qq7dz\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.876843 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-qq7dz\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.877290 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-dns-svc\") pod \"dnsmasq-dns-764c5664d7-qq7dz\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.877671 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-config\") pod \"dnsmasq-dns-764c5664d7-qq7dz\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.878133 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-qq7dz\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.878306 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-qq7dz\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" Nov 24 07:11:17 crc kubenswrapper[4809]: I1124 07:11:17.894071 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4sjs\" (UniqueName: \"kubernetes.io/projected/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-kube-api-access-q4sjs\") pod \"dnsmasq-dns-764c5664d7-qq7dz\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" Nov 24 07:11:18 crc kubenswrapper[4809]: I1124 07:11:18.056753 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" Nov 24 07:11:18 crc kubenswrapper[4809]: I1124 07:11:18.570670 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-qq7dz"] Nov 24 07:11:18 crc kubenswrapper[4809]: W1124 07:11:18.572585 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9de9b687_9a75_41e6_af0e_5dbcf8cbf8a9.slice/crio-bbd6613ca0d114c06e26b5477426bb5e2c5891b3bbd2453cc681601aad859bc5 WatchSource:0}: Error finding container bbd6613ca0d114c06e26b5477426bb5e2c5891b3bbd2453cc681601aad859bc5: Status 404 returned error can't find the container with id bbd6613ca0d114c06e26b5477426bb5e2c5891b3bbd2453cc681601aad859bc5 Nov 24 07:11:18 crc kubenswrapper[4809]: I1124 07:11:18.754118 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vnpwc-config-4nptz" Nov 24 07:11:18 crc kubenswrapper[4809]: I1124 07:11:18.799998 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkpv6\" (UniqueName: \"kubernetes.io/projected/167875f4-19ea-454e-8629-776ba700ac28-kube-api-access-fkpv6\") pod \"167875f4-19ea-454e-8629-776ba700ac28\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " Nov 24 07:11:18 crc kubenswrapper[4809]: I1124 07:11:18.800125 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/167875f4-19ea-454e-8629-776ba700ac28-var-run\") pod \"167875f4-19ea-454e-8629-776ba700ac28\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " Nov 24 07:11:18 crc kubenswrapper[4809]: I1124 07:11:18.800146 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/167875f4-19ea-454e-8629-776ba700ac28-var-log-ovn\") pod \"167875f4-19ea-454e-8629-776ba700ac28\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " Nov 24 07:11:18 crc kubenswrapper[4809]: I1124 07:11:18.800172 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/167875f4-19ea-454e-8629-776ba700ac28-additional-scripts\") pod \"167875f4-19ea-454e-8629-776ba700ac28\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " Nov 24 07:11:18 crc kubenswrapper[4809]: I1124 07:11:18.800203 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/167875f4-19ea-454e-8629-776ba700ac28-var-run-ovn\") pod \"167875f4-19ea-454e-8629-776ba700ac28\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " Nov 24 07:11:18 crc kubenswrapper[4809]: I1124 07:11:18.800246 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/167875f4-19ea-454e-8629-776ba700ac28-scripts\") pod \"167875f4-19ea-454e-8629-776ba700ac28\" (UID: \"167875f4-19ea-454e-8629-776ba700ac28\") " Nov 24 07:11:18 crc kubenswrapper[4809]: I1124 07:11:18.800702 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/167875f4-19ea-454e-8629-776ba700ac28-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "167875f4-19ea-454e-8629-776ba700ac28" (UID: "167875f4-19ea-454e-8629-776ba700ac28"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:11:18 crc kubenswrapper[4809]: I1124 07:11:18.801250 4809 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/167875f4-19ea-454e-8629-776ba700ac28-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:18 crc kubenswrapper[4809]: I1124 07:11:18.801331 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/167875f4-19ea-454e-8629-776ba700ac28-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "167875f4-19ea-454e-8629-776ba700ac28" (UID: "167875f4-19ea-454e-8629-776ba700ac28"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:11:18 crc kubenswrapper[4809]: I1124 07:11:18.801855 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/167875f4-19ea-454e-8629-776ba700ac28-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "167875f4-19ea-454e-8629-776ba700ac28" (UID: "167875f4-19ea-454e-8629-776ba700ac28"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:18 crc kubenswrapper[4809]: I1124 07:11:18.804099 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/167875f4-19ea-454e-8629-776ba700ac28-scripts" (OuterVolumeSpecName: "scripts") pod "167875f4-19ea-454e-8629-776ba700ac28" (UID: "167875f4-19ea-454e-8629-776ba700ac28"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:18 crc kubenswrapper[4809]: I1124 07:11:18.801591 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/167875f4-19ea-454e-8629-776ba700ac28-var-run" (OuterVolumeSpecName: "var-run") pod "167875f4-19ea-454e-8629-776ba700ac28" (UID: "167875f4-19ea-454e-8629-776ba700ac28"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:11:18 crc kubenswrapper[4809]: I1124 07:11:18.807432 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/167875f4-19ea-454e-8629-776ba700ac28-kube-api-access-fkpv6" (OuterVolumeSpecName: "kube-api-access-fkpv6") pod "167875f4-19ea-454e-8629-776ba700ac28" (UID: "167875f4-19ea-454e-8629-776ba700ac28"). InnerVolumeSpecName "kube-api-access-fkpv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:11:18 crc kubenswrapper[4809]: I1124 07:11:18.902744 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkpv6\" (UniqueName: \"kubernetes.io/projected/167875f4-19ea-454e-8629-776ba700ac28-kube-api-access-fkpv6\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:18 crc kubenswrapper[4809]: I1124 07:11:18.903074 4809 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/167875f4-19ea-454e-8629-776ba700ac28-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:18 crc kubenswrapper[4809]: I1124 07:11:18.903084 4809 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/167875f4-19ea-454e-8629-776ba700ac28-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:18 crc kubenswrapper[4809]: I1124 07:11:18.903094 4809 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/167875f4-19ea-454e-8629-776ba700ac28-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:18 crc kubenswrapper[4809]: I1124 07:11:18.903105 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/167875f4-19ea-454e-8629-776ba700ac28-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:19 crc kubenswrapper[4809]: I1124 07:11:19.457473 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vnpwc-config-4nptz" event={"ID":"167875f4-19ea-454e-8629-776ba700ac28","Type":"ContainerDied","Data":"3723685157d353ae927f542c0fe035efeb25a4dbc28a00bf1bec1397700f0f70"} Nov 24 07:11:19 crc kubenswrapper[4809]: I1124 07:11:19.457519 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3723685157d353ae927f542c0fe035efeb25a4dbc28a00bf1bec1397700f0f70" Nov 24 07:11:19 crc kubenswrapper[4809]: I1124 07:11:19.457525 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vnpwc-config-4nptz" Nov 24 07:11:19 crc kubenswrapper[4809]: I1124 07:11:19.459570 4809 generic.go:334] "Generic (PLEG): container finished" podID="9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9" containerID="9ff16c928aec7804af2a3be817ebe72b19fa24a6283012f1d67617e3db0ae2b6" exitCode=0 Nov 24 07:11:19 crc kubenswrapper[4809]: I1124 07:11:19.459608 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" event={"ID":"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9","Type":"ContainerDied","Data":"9ff16c928aec7804af2a3be817ebe72b19fa24a6283012f1d67617e3db0ae2b6"} Nov 24 07:11:19 crc kubenswrapper[4809]: I1124 07:11:19.459642 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" event={"ID":"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9","Type":"ContainerStarted","Data":"bbd6613ca0d114c06e26b5477426bb5e2c5891b3bbd2453cc681601aad859bc5"} Nov 24 07:11:19 crc kubenswrapper[4809]: I1124 07:11:19.823459 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-vnpwc-config-4nptz"] Nov 24 07:11:19 crc kubenswrapper[4809]: I1124 07:11:19.832316 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-vnpwc-config-4nptz"] Nov 24 07:11:20 crc kubenswrapper[4809]: I1124 07:11:20.467497 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" event={"ID":"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9","Type":"ContainerStarted","Data":"cf38cf8da3467a17cabe18059f318c623c5d9e13bfcc3aa5bf5ccf71e6e9e479"} Nov 24 07:11:20 crc kubenswrapper[4809]: I1124 07:11:20.467632 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" Nov 24 07:11:20 crc kubenswrapper[4809]: I1124 07:11:20.469861 4809 generic.go:334] "Generic (PLEG): container finished" podID="952df16f-264d-49ca-9281-9f3b95f8273d" containerID="9e6f28cb3c7b8bc7f9374a06b3bc896f2892963f6b563dd75c5d836ebfb60db8" exitCode=0 Nov 24 07:11:20 crc kubenswrapper[4809]: I1124 07:11:20.469892 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-lfw7d" event={"ID":"952df16f-264d-49ca-9281-9f3b95f8273d","Type":"ContainerDied","Data":"9e6f28cb3c7b8bc7f9374a06b3bc896f2892963f6b563dd75c5d836ebfb60db8"} Nov 24 07:11:20 crc kubenswrapper[4809]: I1124 07:11:20.488573 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" podStartSLOduration=3.488549832 podStartE2EDuration="3.488549832s" podCreationTimestamp="2025-11-24 07:11:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:11:20.486657802 +0000 UTC m=+980.387249517" watchObservedRunningTime="2025-11-24 07:11:20.488549832 +0000 UTC m=+980.389141557" Nov 24 07:11:20 crc kubenswrapper[4809]: I1124 07:11:20.904306 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="167875f4-19ea-454e-8629-776ba700ac28" path="/var/lib/kubelet/pods/167875f4-19ea-454e-8629-776ba700ac28/volumes" Nov 24 07:11:21 crc kubenswrapper[4809]: I1124 07:11:21.821954 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-lfw7d" Nov 24 07:11:21 crc kubenswrapper[4809]: I1124 07:11:21.847633 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/952df16f-264d-49ca-9281-9f3b95f8273d-db-sync-config-data\") pod \"952df16f-264d-49ca-9281-9f3b95f8273d\" (UID: \"952df16f-264d-49ca-9281-9f3b95f8273d\") " Nov 24 07:11:21 crc kubenswrapper[4809]: I1124 07:11:21.847775 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/952df16f-264d-49ca-9281-9f3b95f8273d-config-data\") pod \"952df16f-264d-49ca-9281-9f3b95f8273d\" (UID: \"952df16f-264d-49ca-9281-9f3b95f8273d\") " Nov 24 07:11:21 crc kubenswrapper[4809]: I1124 07:11:21.847814 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtm8d\" (UniqueName: \"kubernetes.io/projected/952df16f-264d-49ca-9281-9f3b95f8273d-kube-api-access-wtm8d\") pod \"952df16f-264d-49ca-9281-9f3b95f8273d\" (UID: \"952df16f-264d-49ca-9281-9f3b95f8273d\") " Nov 24 07:11:21 crc kubenswrapper[4809]: I1124 07:11:21.847879 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/952df16f-264d-49ca-9281-9f3b95f8273d-combined-ca-bundle\") pod \"952df16f-264d-49ca-9281-9f3b95f8273d\" (UID: \"952df16f-264d-49ca-9281-9f3b95f8273d\") " Nov 24 07:11:21 crc kubenswrapper[4809]: I1124 07:11:21.855010 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/952df16f-264d-49ca-9281-9f3b95f8273d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "952df16f-264d-49ca-9281-9f3b95f8273d" (UID: "952df16f-264d-49ca-9281-9f3b95f8273d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:11:21 crc kubenswrapper[4809]: I1124 07:11:21.855096 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/952df16f-264d-49ca-9281-9f3b95f8273d-kube-api-access-wtm8d" (OuterVolumeSpecName: "kube-api-access-wtm8d") pod "952df16f-264d-49ca-9281-9f3b95f8273d" (UID: "952df16f-264d-49ca-9281-9f3b95f8273d"). InnerVolumeSpecName "kube-api-access-wtm8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:11:21 crc kubenswrapper[4809]: I1124 07:11:21.881793 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/952df16f-264d-49ca-9281-9f3b95f8273d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "952df16f-264d-49ca-9281-9f3b95f8273d" (UID: "952df16f-264d-49ca-9281-9f3b95f8273d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:11:21 crc kubenswrapper[4809]: I1124 07:11:21.899360 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/952df16f-264d-49ca-9281-9f3b95f8273d-config-data" (OuterVolumeSpecName: "config-data") pod "952df16f-264d-49ca-9281-9f3b95f8273d" (UID: "952df16f-264d-49ca-9281-9f3b95f8273d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:11:21 crc kubenswrapper[4809]: I1124 07:11:21.949713 4809 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/952df16f-264d-49ca-9281-9f3b95f8273d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:21 crc kubenswrapper[4809]: I1124 07:11:21.949768 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/952df16f-264d-49ca-9281-9f3b95f8273d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:21 crc kubenswrapper[4809]: I1124 07:11:21.949786 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtm8d\" (UniqueName: \"kubernetes.io/projected/952df16f-264d-49ca-9281-9f3b95f8273d-kube-api-access-wtm8d\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:21 crc kubenswrapper[4809]: I1124 07:11:21.949804 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/952df16f-264d-49ca-9281-9f3b95f8273d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:22 crc kubenswrapper[4809]: I1124 07:11:22.471316 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 07:11:22 crc kubenswrapper[4809]: I1124 07:11:22.495845 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-lfw7d" event={"ID":"952df16f-264d-49ca-9281-9f3b95f8273d","Type":"ContainerDied","Data":"31ff16a55d2882fb1dbce8d1349259cbe85d92224b284156c87db82aad4d2182"} Nov 24 07:11:22 crc kubenswrapper[4809]: I1124 07:11:22.495880 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31ff16a55d2882fb1dbce8d1349259cbe85d92224b284156c87db82aad4d2182" Nov 24 07:11:22 crc kubenswrapper[4809]: I1124 07:11:22.495932 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-lfw7d" Nov 24 07:11:22 crc kubenswrapper[4809]: I1124 07:11:22.557162 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:11:22 crc kubenswrapper[4809]: I1124 07:11:22.801539 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-vxq56"] Nov 24 07:11:22 crc kubenswrapper[4809]: E1124 07:11:22.801913 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="167875f4-19ea-454e-8629-776ba700ac28" containerName="ovn-config" Nov 24 07:11:22 crc kubenswrapper[4809]: I1124 07:11:22.801940 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="167875f4-19ea-454e-8629-776ba700ac28" containerName="ovn-config" Nov 24 07:11:22 crc kubenswrapper[4809]: E1124 07:11:22.801953 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="952df16f-264d-49ca-9281-9f3b95f8273d" containerName="glance-db-sync" Nov 24 07:11:22 crc kubenswrapper[4809]: I1124 07:11:22.801975 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="952df16f-264d-49ca-9281-9f3b95f8273d" containerName="glance-db-sync" Nov 24 07:11:22 crc kubenswrapper[4809]: I1124 07:11:22.802161 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="952df16f-264d-49ca-9281-9f3b95f8273d" containerName="glance-db-sync" Nov 24 07:11:22 crc kubenswrapper[4809]: I1124 07:11:22.802174 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="167875f4-19ea-454e-8629-776ba700ac28" containerName="ovn-config" Nov 24 07:11:22 crc kubenswrapper[4809]: I1124 07:11:22.802788 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vxq56" Nov 24 07:11:22 crc kubenswrapper[4809]: I1124 07:11:22.834997 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vxq56"] Nov 24 07:11:22 crc kubenswrapper[4809]: I1124 07:11:22.864743 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4e0697a-3b6b-4d9c-a863-9ef8a696920a-operator-scripts\") pod \"cinder-db-create-vxq56\" (UID: \"f4e0697a-3b6b-4d9c-a863-9ef8a696920a\") " pod="openstack/cinder-db-create-vxq56" Nov 24 07:11:22 crc kubenswrapper[4809]: I1124 07:11:22.864858 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5bvj\" (UniqueName: \"kubernetes.io/projected/f4e0697a-3b6b-4d9c-a863-9ef8a696920a-kube-api-access-t5bvj\") pod \"cinder-db-create-vxq56\" (UID: \"f4e0697a-3b6b-4d9c-a863-9ef8a696920a\") " pod="openstack/cinder-db-create-vxq56" Nov 24 07:11:22 crc kubenswrapper[4809]: I1124 07:11:22.966425 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4e0697a-3b6b-4d9c-a863-9ef8a696920a-operator-scripts\") pod \"cinder-db-create-vxq56\" (UID: \"f4e0697a-3b6b-4d9c-a863-9ef8a696920a\") " pod="openstack/cinder-db-create-vxq56" Nov 24 07:11:22 crc kubenswrapper[4809]: I1124 07:11:22.966729 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5bvj\" (UniqueName: \"kubernetes.io/projected/f4e0697a-3b6b-4d9c-a863-9ef8a696920a-kube-api-access-t5bvj\") pod \"cinder-db-create-vxq56\" (UID: \"f4e0697a-3b6b-4d9c-a863-9ef8a696920a\") " pod="openstack/cinder-db-create-vxq56" Nov 24 07:11:22 crc kubenswrapper[4809]: I1124 07:11:22.967102 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4e0697a-3b6b-4d9c-a863-9ef8a696920a-operator-scripts\") pod \"cinder-db-create-vxq56\" (UID: \"f4e0697a-3b6b-4d9c-a863-9ef8a696920a\") " pod="openstack/cinder-db-create-vxq56" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.007087 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5bvj\" (UniqueName: \"kubernetes.io/projected/f4e0697a-3b6b-4d9c-a863-9ef8a696920a-kube-api-access-t5bvj\") pod \"cinder-db-create-vxq56\" (UID: \"f4e0697a-3b6b-4d9c-a863-9ef8a696920a\") " pod="openstack/cinder-db-create-vxq56" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.012701 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-00f1-account-create-pfn7s"] Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.013701 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-00f1-account-create-pfn7s" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.034923 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.045328 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-dnl8c"] Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.046538 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dnl8c" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.052041 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-00f1-account-create-pfn7s"] Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.071725 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-dnl8c"] Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.119295 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vxq56" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.170797 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsdxj\" (UniqueName: \"kubernetes.io/projected/7bcc82f1-7d2a-48b5-83cb-157f3d4c320c-kube-api-access-wsdxj\") pod \"cinder-00f1-account-create-pfn7s\" (UID: \"7bcc82f1-7d2a-48b5-83cb-157f3d4c320c\") " pod="openstack/cinder-00f1-account-create-pfn7s" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.170867 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbkf9\" (UniqueName: \"kubernetes.io/projected/1bad348b-3d85-4a22-9d03-734715eedffa-kube-api-access-dbkf9\") pod \"barbican-db-create-dnl8c\" (UID: \"1bad348b-3d85-4a22-9d03-734715eedffa\") " pod="openstack/barbican-db-create-dnl8c" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.170914 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bcc82f1-7d2a-48b5-83cb-157f3d4c320c-operator-scripts\") pod \"cinder-00f1-account-create-pfn7s\" (UID: \"7bcc82f1-7d2a-48b5-83cb-157f3d4c320c\") " pod="openstack/cinder-00f1-account-create-pfn7s" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.171046 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1bad348b-3d85-4a22-9d03-734715eedffa-operator-scripts\") pod \"barbican-db-create-dnl8c\" (UID: \"1bad348b-3d85-4a22-9d03-734715eedffa\") " pod="openstack/barbican-db-create-dnl8c" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.244363 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-2540-account-create-w7dp4"] Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.245344 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2540-account-create-w7dp4" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.260146 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.272563 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsdxj\" (UniqueName: \"kubernetes.io/projected/7bcc82f1-7d2a-48b5-83cb-157f3d4c320c-kube-api-access-wsdxj\") pod \"cinder-00f1-account-create-pfn7s\" (UID: \"7bcc82f1-7d2a-48b5-83cb-157f3d4c320c\") " pod="openstack/cinder-00f1-account-create-pfn7s" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.272630 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbkf9\" (UniqueName: \"kubernetes.io/projected/1bad348b-3d85-4a22-9d03-734715eedffa-kube-api-access-dbkf9\") pod \"barbican-db-create-dnl8c\" (UID: \"1bad348b-3d85-4a22-9d03-734715eedffa\") " pod="openstack/barbican-db-create-dnl8c" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.272678 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bcc82f1-7d2a-48b5-83cb-157f3d4c320c-operator-scripts\") pod \"cinder-00f1-account-create-pfn7s\" (UID: \"7bcc82f1-7d2a-48b5-83cb-157f3d4c320c\") " pod="openstack/cinder-00f1-account-create-pfn7s" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.272749 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1bad348b-3d85-4a22-9d03-734715eedffa-operator-scripts\") pod \"barbican-db-create-dnl8c\" (UID: \"1bad348b-3d85-4a22-9d03-734715eedffa\") " pod="openstack/barbican-db-create-dnl8c" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.274304 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bcc82f1-7d2a-48b5-83cb-157f3d4c320c-operator-scripts\") pod \"cinder-00f1-account-create-pfn7s\" (UID: \"7bcc82f1-7d2a-48b5-83cb-157f3d4c320c\") " pod="openstack/cinder-00f1-account-create-pfn7s" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.274337 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1bad348b-3d85-4a22-9d03-734715eedffa-operator-scripts\") pod \"barbican-db-create-dnl8c\" (UID: \"1bad348b-3d85-4a22-9d03-734715eedffa\") " pod="openstack/barbican-db-create-dnl8c" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.276324 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-qq7dz"] Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.276524 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" podUID="9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9" containerName="dnsmasq-dns" containerID="cri-o://cf38cf8da3467a17cabe18059f318c623c5d9e13bfcc3aa5bf5ccf71e6e9e479" gracePeriod=10 Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.333837 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-2540-account-create-w7dp4"] Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.339425 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsdxj\" (UniqueName: \"kubernetes.io/projected/7bcc82f1-7d2a-48b5-83cb-157f3d4c320c-kube-api-access-wsdxj\") pod \"cinder-00f1-account-create-pfn7s\" (UID: \"7bcc82f1-7d2a-48b5-83cb-157f3d4c320c\") " pod="openstack/cinder-00f1-account-create-pfn7s" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.349557 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbkf9\" (UniqueName: \"kubernetes.io/projected/1bad348b-3d85-4a22-9d03-734715eedffa-kube-api-access-dbkf9\") pod \"barbican-db-create-dnl8c\" (UID: \"1bad348b-3d85-4a22-9d03-734715eedffa\") " pod="openstack/barbican-db-create-dnl8c" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.358492 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-00f1-account-create-pfn7s" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.371604 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dnl8c" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.376182 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06f78e27-776f-4506-b3f2-a348318cbd13-operator-scripts\") pod \"barbican-2540-account-create-w7dp4\" (UID: \"06f78e27-776f-4506-b3f2-a348318cbd13\") " pod="openstack/barbican-2540-account-create-w7dp4" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.376332 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7j586\" (UniqueName: \"kubernetes.io/projected/06f78e27-776f-4506-b3f2-a348318cbd13-kube-api-access-7j586\") pod \"barbican-2540-account-create-w7dp4\" (UID: \"06f78e27-776f-4506-b3f2-a348318cbd13\") " pod="openstack/barbican-2540-account-create-w7dp4" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.382032 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-2b7fh"] Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.394857 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.409941 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-2b7fh"] Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.467747 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-tbhvc"] Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.472961 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-tbhvc" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.477756 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06f78e27-776f-4506-b3f2-a348318cbd13-operator-scripts\") pod \"barbican-2540-account-create-w7dp4\" (UID: \"06f78e27-776f-4506-b3f2-a348318cbd13\") " pod="openstack/barbican-2540-account-create-w7dp4" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.477843 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7j586\" (UniqueName: \"kubernetes.io/projected/06f78e27-776f-4506-b3f2-a348318cbd13-kube-api-access-7j586\") pod \"barbican-2540-account-create-w7dp4\" (UID: \"06f78e27-776f-4506-b3f2-a348318cbd13\") " pod="openstack/barbican-2540-account-create-w7dp4" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.478806 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06f78e27-776f-4506-b3f2-a348318cbd13-operator-scripts\") pod \"barbican-2540-account-create-w7dp4\" (UID: \"06f78e27-776f-4506-b3f2-a348318cbd13\") " pod="openstack/barbican-2540-account-create-w7dp4" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.479149 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.480889 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.482391 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fkz7g" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.482557 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.494241 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-pz5bj"] Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.495419 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pz5bj" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.506467 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-tbhvc"] Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.515947 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7j586\" (UniqueName: \"kubernetes.io/projected/06f78e27-776f-4506-b3f2-a348318cbd13-kube-api-access-7j586\") pod \"barbican-2540-account-create-w7dp4\" (UID: \"06f78e27-776f-4506-b3f2-a348318cbd13\") " pod="openstack/barbican-2540-account-create-w7dp4" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.545366 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-pz5bj"] Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.548254 4809 generic.go:334] "Generic (PLEG): container finished" podID="9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9" containerID="cf38cf8da3467a17cabe18059f318c623c5d9e13bfcc3aa5bf5ccf71e6e9e479" exitCode=0 Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.548304 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" event={"ID":"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9","Type":"ContainerDied","Data":"cf38cf8da3467a17cabe18059f318c623c5d9e13bfcc3aa5bf5ccf71e6e9e479"} Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.549884 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-e4c1-account-create-4f5hg"] Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.551692 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e4c1-account-create-4f5hg" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.553888 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.564945 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-e4c1-account-create-4f5hg"] Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.576338 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2540-account-create-w7dp4" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.580840 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99c73ccf-06f4-4dfb-ab11-a1b93ab653ea-config-data\") pod \"keystone-db-sync-tbhvc\" (UID: \"99c73ccf-06f4-4dfb-ab11-a1b93ab653ea\") " pod="openstack/keystone-db-sync-tbhvc" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.581039 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-2b7fh\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.581064 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99c73ccf-06f4-4dfb-ab11-a1b93ab653ea-combined-ca-bundle\") pod \"keystone-db-sync-tbhvc\" (UID: \"99c73ccf-06f4-4dfb-ab11-a1b93ab653ea\") " pod="openstack/keystone-db-sync-tbhvc" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.581084 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-2b7fh\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.581108 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7q6t\" (UniqueName: \"kubernetes.io/projected/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-kube-api-access-l7q6t\") pod \"dnsmasq-dns-74f6bcbc87-2b7fh\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.581124 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-2b7fh\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.581142 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-config\") pod \"dnsmasq-dns-74f6bcbc87-2b7fh\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.581164 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6nhk\" (UniqueName: \"kubernetes.io/projected/99c73ccf-06f4-4dfb-ab11-a1b93ab653ea-kube-api-access-x6nhk\") pod \"keystone-db-sync-tbhvc\" (UID: \"99c73ccf-06f4-4dfb-ab11-a1b93ab653ea\") " pod="openstack/keystone-db-sync-tbhvc" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.581192 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-2b7fh\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.682273 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7q6t\" (UniqueName: \"kubernetes.io/projected/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-kube-api-access-l7q6t\") pod \"dnsmasq-dns-74f6bcbc87-2b7fh\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.682533 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-2b7fh\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.682554 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91db574f-2e79-4e56-a408-64373360aa4a-operator-scripts\") pod \"neutron-db-create-pz5bj\" (UID: \"91db574f-2e79-4e56-a408-64373360aa4a\") " pod="openstack/neutron-db-create-pz5bj" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.682582 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-config\") pod \"dnsmasq-dns-74f6bcbc87-2b7fh\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.682608 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6nhk\" (UniqueName: \"kubernetes.io/projected/99c73ccf-06f4-4dfb-ab11-a1b93ab653ea-kube-api-access-x6nhk\") pod \"keystone-db-sync-tbhvc\" (UID: \"99c73ccf-06f4-4dfb-ab11-a1b93ab653ea\") " pod="openstack/keystone-db-sync-tbhvc" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.682645 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-2b7fh\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.682677 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2np6\" (UniqueName: \"kubernetes.io/projected/7c34951d-7bde-4c8e-b369-fc6505c3a752-kube-api-access-m2np6\") pod \"neutron-e4c1-account-create-4f5hg\" (UID: \"7c34951d-7bde-4c8e-b369-fc6505c3a752\") " pod="openstack/neutron-e4c1-account-create-4f5hg" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.682691 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf9cv\" (UniqueName: \"kubernetes.io/projected/91db574f-2e79-4e56-a408-64373360aa4a-kube-api-access-lf9cv\") pod \"neutron-db-create-pz5bj\" (UID: \"91db574f-2e79-4e56-a408-64373360aa4a\") " pod="openstack/neutron-db-create-pz5bj" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.682711 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99c73ccf-06f4-4dfb-ab11-a1b93ab653ea-config-data\") pod \"keystone-db-sync-tbhvc\" (UID: \"99c73ccf-06f4-4dfb-ab11-a1b93ab653ea\") " pod="openstack/keystone-db-sync-tbhvc" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.682765 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c34951d-7bde-4c8e-b369-fc6505c3a752-operator-scripts\") pod \"neutron-e4c1-account-create-4f5hg\" (UID: \"7c34951d-7bde-4c8e-b369-fc6505c3a752\") " pod="openstack/neutron-e4c1-account-create-4f5hg" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.682808 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-2b7fh\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.682825 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99c73ccf-06f4-4dfb-ab11-a1b93ab653ea-combined-ca-bundle\") pod \"keystone-db-sync-tbhvc\" (UID: \"99c73ccf-06f4-4dfb-ab11-a1b93ab653ea\") " pod="openstack/keystone-db-sync-tbhvc" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.682839 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-2b7fh\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.683781 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-2b7fh\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.684518 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-2b7fh\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.684832 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-2b7fh\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.685201 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-config\") pod \"dnsmasq-dns-74f6bcbc87-2b7fh\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.685523 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-2b7fh\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.691087 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99c73ccf-06f4-4dfb-ab11-a1b93ab653ea-config-data\") pod \"keystone-db-sync-tbhvc\" (UID: \"99c73ccf-06f4-4dfb-ab11-a1b93ab653ea\") " pod="openstack/keystone-db-sync-tbhvc" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.694434 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99c73ccf-06f4-4dfb-ab11-a1b93ab653ea-combined-ca-bundle\") pod \"keystone-db-sync-tbhvc\" (UID: \"99c73ccf-06f4-4dfb-ab11-a1b93ab653ea\") " pod="openstack/keystone-db-sync-tbhvc" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.707472 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6nhk\" (UniqueName: \"kubernetes.io/projected/99c73ccf-06f4-4dfb-ab11-a1b93ab653ea-kube-api-access-x6nhk\") pod \"keystone-db-sync-tbhvc\" (UID: \"99c73ccf-06f4-4dfb-ab11-a1b93ab653ea\") " pod="openstack/keystone-db-sync-tbhvc" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.711686 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7q6t\" (UniqueName: \"kubernetes.io/projected/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-kube-api-access-l7q6t\") pod \"dnsmasq-dns-74f6bcbc87-2b7fh\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.783788 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91db574f-2e79-4e56-a408-64373360aa4a-operator-scripts\") pod \"neutron-db-create-pz5bj\" (UID: \"91db574f-2e79-4e56-a408-64373360aa4a\") " pod="openstack/neutron-db-create-pz5bj" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.783913 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2np6\" (UniqueName: \"kubernetes.io/projected/7c34951d-7bde-4c8e-b369-fc6505c3a752-kube-api-access-m2np6\") pod \"neutron-e4c1-account-create-4f5hg\" (UID: \"7c34951d-7bde-4c8e-b369-fc6505c3a752\") " pod="openstack/neutron-e4c1-account-create-4f5hg" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.783938 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf9cv\" (UniqueName: \"kubernetes.io/projected/91db574f-2e79-4e56-a408-64373360aa4a-kube-api-access-lf9cv\") pod \"neutron-db-create-pz5bj\" (UID: \"91db574f-2e79-4e56-a408-64373360aa4a\") " pod="openstack/neutron-db-create-pz5bj" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.784025 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c34951d-7bde-4c8e-b369-fc6505c3a752-operator-scripts\") pod \"neutron-e4c1-account-create-4f5hg\" (UID: \"7c34951d-7bde-4c8e-b369-fc6505c3a752\") " pod="openstack/neutron-e4c1-account-create-4f5hg" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.793116 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91db574f-2e79-4e56-a408-64373360aa4a-operator-scripts\") pod \"neutron-db-create-pz5bj\" (UID: \"91db574f-2e79-4e56-a408-64373360aa4a\") " pod="openstack/neutron-db-create-pz5bj" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.794934 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c34951d-7bde-4c8e-b369-fc6505c3a752-operator-scripts\") pod \"neutron-e4c1-account-create-4f5hg\" (UID: \"7c34951d-7bde-4c8e-b369-fc6505c3a752\") " pod="openstack/neutron-e4c1-account-create-4f5hg" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.801522 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2np6\" (UniqueName: \"kubernetes.io/projected/7c34951d-7bde-4c8e-b369-fc6505c3a752-kube-api-access-m2np6\") pod \"neutron-e4c1-account-create-4f5hg\" (UID: \"7c34951d-7bde-4c8e-b369-fc6505c3a752\") " pod="openstack/neutron-e4c1-account-create-4f5hg" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.806502 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf9cv\" (UniqueName: \"kubernetes.io/projected/91db574f-2e79-4e56-a408-64373360aa4a-kube-api-access-lf9cv\") pod \"neutron-db-create-pz5bj\" (UID: \"91db574f-2e79-4e56-a408-64373360aa4a\") " pod="openstack/neutron-db-create-pz5bj" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.813106 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.837878 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vxq56"] Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.840091 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-tbhvc" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.854063 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pz5bj" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.889875 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-2540-account-create-w7dp4"] Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.893326 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e4c1-account-create-4f5hg" Nov 24 07:11:23 crc kubenswrapper[4809]: I1124 07:11:23.989600 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-00f1-account-create-pfn7s"] Nov 24 07:11:24 crc kubenswrapper[4809]: I1124 07:11:24.141837 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-dnl8c"] Nov 24 07:11:24 crc kubenswrapper[4809]: W1124 07:11:24.156269 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1bad348b_3d85_4a22_9d03_734715eedffa.slice/crio-42f789eaf4b1ef88016a2f89b119e8e6ee5bd74b7a4748a832dfd653c2bd8e1a WatchSource:0}: Error finding container 42f789eaf4b1ef88016a2f89b119e8e6ee5bd74b7a4748a832dfd653c2bd8e1a: Status 404 returned error can't find the container with id 42f789eaf4b1ef88016a2f89b119e8e6ee5bd74b7a4748a832dfd653c2bd8e1a Nov 24 07:11:24 crc kubenswrapper[4809]: I1124 07:11:24.444665 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-2b7fh"] Nov 24 07:11:24 crc kubenswrapper[4809]: W1124 07:11:24.456366 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c30be2a_1a3c_41c2_95fc_1f1ad8e23ed5.slice/crio-0da44b051b38ed2c65e0b9a2118051637203c5bff074f9eb1e37b02528fec3fa WatchSource:0}: Error finding container 0da44b051b38ed2c65e0b9a2118051637203c5bff074f9eb1e37b02528fec3fa: Status 404 returned error can't find the container with id 0da44b051b38ed2c65e0b9a2118051637203c5bff074f9eb1e37b02528fec3fa Nov 24 07:11:24 crc kubenswrapper[4809]: I1124 07:11:24.465784 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-e4c1-account-create-4f5hg"] Nov 24 07:11:24 crc kubenswrapper[4809]: I1124 07:11:24.537925 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-tbhvc"] Nov 24 07:11:24 crc kubenswrapper[4809]: I1124 07:11:24.565162 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-00f1-account-create-pfn7s" event={"ID":"7bcc82f1-7d2a-48b5-83cb-157f3d4c320c","Type":"ContainerStarted","Data":"5416cb49412a03dab5cdf3aa742a55fa985b37149f235bc66f649522b61f855a"} Nov 24 07:11:24 crc kubenswrapper[4809]: I1124 07:11:24.569242 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e4c1-account-create-4f5hg" event={"ID":"7c34951d-7bde-4c8e-b369-fc6505c3a752","Type":"ContainerStarted","Data":"8d223a28d3c8f3e6f5d0e008f9c1d62d5110812c01a9adf6c3e64c0b7b39250a"} Nov 24 07:11:24 crc kubenswrapper[4809]: I1124 07:11:24.570463 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" event={"ID":"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5","Type":"ContainerStarted","Data":"0da44b051b38ed2c65e0b9a2118051637203c5bff074f9eb1e37b02528fec3fa"} Nov 24 07:11:24 crc kubenswrapper[4809]: I1124 07:11:24.572016 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vxq56" event={"ID":"f4e0697a-3b6b-4d9c-a863-9ef8a696920a","Type":"ContainerStarted","Data":"bdc4863e5661c54e551bb8fc869ec484f3d8c028fff26e4c637a1a0914b342c3"} Nov 24 07:11:24 crc kubenswrapper[4809]: I1124 07:11:24.573420 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2540-account-create-w7dp4" event={"ID":"06f78e27-776f-4506-b3f2-a348318cbd13","Type":"ContainerStarted","Data":"7951014b50694cb7e4a047e7d5ac204653bb556d9c6caccff9d91d20a4b8677d"} Nov 24 07:11:24 crc kubenswrapper[4809]: I1124 07:11:24.576058 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dnl8c" event={"ID":"1bad348b-3d85-4a22-9d03-734715eedffa","Type":"ContainerStarted","Data":"42f789eaf4b1ef88016a2f89b119e8e6ee5bd74b7a4748a832dfd653c2bd8e1a"} Nov 24 07:11:24 crc kubenswrapper[4809]: I1124 07:11:24.584442 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-pz5bj"] Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.557093 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.617473 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-tbhvc" event={"ID":"99c73ccf-06f4-4dfb-ab11-a1b93ab653ea","Type":"ContainerStarted","Data":"7a83c2085a05d8fdff741cd0d3f85f26c7ed75eae5196fb8599694affbdef865"} Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.622923 4809 generic.go:334] "Generic (PLEG): container finished" podID="06f78e27-776f-4506-b3f2-a348318cbd13" containerID="d974c3beaef64ba1537b2ec4e909be97aded61fe58f8b13cb2d636c170100f74" exitCode=0 Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.622989 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2540-account-create-w7dp4" event={"ID":"06f78e27-776f-4506-b3f2-a348318cbd13","Type":"ContainerDied","Data":"d974c3beaef64ba1537b2ec4e909be97aded61fe58f8b13cb2d636c170100f74"} Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.627176 4809 generic.go:334] "Generic (PLEG): container finished" podID="1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5" containerID="056e7b7f88d76dbaaa1c9ceaa438598701da283180eba591690d9e0dba040281" exitCode=0 Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.627255 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" event={"ID":"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5","Type":"ContainerDied","Data":"056e7b7f88d76dbaaa1c9ceaa438598701da283180eba591690d9e0dba040281"} Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.632492 4809 generic.go:334] "Generic (PLEG): container finished" podID="1bad348b-3d85-4a22-9d03-734715eedffa" containerID="98853418d288a056113d9c613919ff0f6958064a0058aeaff0e497d4a507db03" exitCode=0 Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.632565 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dnl8c" event={"ID":"1bad348b-3d85-4a22-9d03-734715eedffa","Type":"ContainerDied","Data":"98853418d288a056113d9c613919ff0f6958064a0058aeaff0e497d4a507db03"} Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.636348 4809 generic.go:334] "Generic (PLEG): container finished" podID="91db574f-2e79-4e56-a408-64373360aa4a" containerID="3fbf20b7ca69e9111cab29ca1b274c354c8ab7644e8fb693ff3bc3e0d58658cd" exitCode=0 Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.636758 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pz5bj" event={"ID":"91db574f-2e79-4e56-a408-64373360aa4a","Type":"ContainerDied","Data":"3fbf20b7ca69e9111cab29ca1b274c354c8ab7644e8fb693ff3bc3e0d58658cd"} Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.636813 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pz5bj" event={"ID":"91db574f-2e79-4e56-a408-64373360aa4a","Type":"ContainerStarted","Data":"3f8d085ad57d887407403b371bb7ef758a4dae964924809efd3facff9de4bbfe"} Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.642389 4809 generic.go:334] "Generic (PLEG): container finished" podID="7bcc82f1-7d2a-48b5-83cb-157f3d4c320c" containerID="6e49c55d9b850ae38097aee946cdd838022b3d5f6af1325c50a9b6038bb71579" exitCode=0 Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.642547 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-00f1-account-create-pfn7s" event={"ID":"7bcc82f1-7d2a-48b5-83cb-157f3d4c320c","Type":"ContainerDied","Data":"6e49c55d9b850ae38097aee946cdd838022b3d5f6af1325c50a9b6038bb71579"} Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.646862 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" event={"ID":"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9","Type":"ContainerDied","Data":"bbd6613ca0d114c06e26b5477426bb5e2c5891b3bbd2453cc681601aad859bc5"} Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.647109 4809 scope.go:117] "RemoveContainer" containerID="cf38cf8da3467a17cabe18059f318c623c5d9e13bfcc3aa5bf5ccf71e6e9e479" Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.647135 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-qq7dz" Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.648474 4809 generic.go:334] "Generic (PLEG): container finished" podID="7c34951d-7bde-4c8e-b369-fc6505c3a752" containerID="56318ea24cf9f8d2440b2c2e6c2eba928409aeac48663b72a34e11fd39e3dff6" exitCode=0 Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.648546 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e4c1-account-create-4f5hg" event={"ID":"7c34951d-7bde-4c8e-b369-fc6505c3a752","Type":"ContainerDied","Data":"56318ea24cf9f8d2440b2c2e6c2eba928409aeac48663b72a34e11fd39e3dff6"} Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.650273 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4e0697a-3b6b-4d9c-a863-9ef8a696920a" containerID="7585ffb5da77f70ac97e781e5b676cee5a9d97f3023594a6cd2d14a0f2bf4bac" exitCode=0 Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.650305 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vxq56" event={"ID":"f4e0697a-3b6b-4d9c-a863-9ef8a696920a","Type":"ContainerDied","Data":"7585ffb5da77f70ac97e781e5b676cee5a9d97f3023594a6cd2d14a0f2bf4bac"} Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.716127 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-ovsdbserver-sb\") pod \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.728064 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-kube-api-access-q4sjs" (OuterVolumeSpecName: "kube-api-access-q4sjs") pod "9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9" (UID: "9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9"). InnerVolumeSpecName "kube-api-access-q4sjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.728220 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4sjs\" (UniqueName: \"kubernetes.io/projected/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-kube-api-access-q4sjs\") pod \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.728381 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-dns-swift-storage-0\") pod \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.728469 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-config\") pod \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.728534 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-ovsdbserver-nb\") pod \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.728588 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-dns-svc\") pod \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\" (UID: \"9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9\") " Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.750542 4809 scope.go:117] "RemoveContainer" containerID="9ff16c928aec7804af2a3be817ebe72b19fa24a6283012f1d67617e3db0ae2b6" Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.808752 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9" (UID: "9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.812161 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9" (UID: "9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.828710 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9" (UID: "9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.829639 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-config" (OuterVolumeSpecName: "config") pod "9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9" (UID: "9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.830166 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4sjs\" (UniqueName: \"kubernetes.io/projected/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-kube-api-access-q4sjs\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.830187 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.830201 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.830213 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.830255 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.847827 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9" (UID: "9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.932184 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.987434 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-qq7dz"] Nov 24 07:11:25 crc kubenswrapper[4809]: I1124 07:11:25.993156 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-qq7dz"] Nov 24 07:11:26 crc kubenswrapper[4809]: I1124 07:11:26.666839 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" event={"ID":"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5","Type":"ContainerStarted","Data":"95db14eaac6f10051b603ed6695d3a7360907b8e405a445d744910d120259bbe"} Nov 24 07:11:26 crc kubenswrapper[4809]: I1124 07:11:26.667294 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:11:26 crc kubenswrapper[4809]: I1124 07:11:26.694221 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" podStartSLOduration=3.694198535 podStartE2EDuration="3.694198535s" podCreationTimestamp="2025-11-24 07:11:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:11:26.691797762 +0000 UTC m=+986.592389467" watchObservedRunningTime="2025-11-24 07:11:26.694198535 +0000 UTC m=+986.594790250" Nov 24 07:11:26 crc kubenswrapper[4809]: I1124 07:11:26.905447 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9" path="/var/lib/kubelet/pods/9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9/volumes" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.564246 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dnl8c" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.573042 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-00f1-account-create-pfn7s" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.598513 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2540-account-create-w7dp4" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.602733 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsdxj\" (UniqueName: \"kubernetes.io/projected/7bcc82f1-7d2a-48b5-83cb-157f3d4c320c-kube-api-access-wsdxj\") pod \"7bcc82f1-7d2a-48b5-83cb-157f3d4c320c\" (UID: \"7bcc82f1-7d2a-48b5-83cb-157f3d4c320c\") " Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.602774 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bcc82f1-7d2a-48b5-83cb-157f3d4c320c-operator-scripts\") pod \"7bcc82f1-7d2a-48b5-83cb-157f3d4c320c\" (UID: \"7bcc82f1-7d2a-48b5-83cb-157f3d4c320c\") " Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.602797 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbkf9\" (UniqueName: \"kubernetes.io/projected/1bad348b-3d85-4a22-9d03-734715eedffa-kube-api-access-dbkf9\") pod \"1bad348b-3d85-4a22-9d03-734715eedffa\" (UID: \"1bad348b-3d85-4a22-9d03-734715eedffa\") " Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.602829 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1bad348b-3d85-4a22-9d03-734715eedffa-operator-scripts\") pod \"1bad348b-3d85-4a22-9d03-734715eedffa\" (UID: \"1bad348b-3d85-4a22-9d03-734715eedffa\") " Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.608590 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bad348b-3d85-4a22-9d03-734715eedffa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1bad348b-3d85-4a22-9d03-734715eedffa" (UID: "1bad348b-3d85-4a22-9d03-734715eedffa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.609352 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bcc82f1-7d2a-48b5-83cb-157f3d4c320c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7bcc82f1-7d2a-48b5-83cb-157f3d4c320c" (UID: "7bcc82f1-7d2a-48b5-83cb-157f3d4c320c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.612727 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bad348b-3d85-4a22-9d03-734715eedffa-kube-api-access-dbkf9" (OuterVolumeSpecName: "kube-api-access-dbkf9") pod "1bad348b-3d85-4a22-9d03-734715eedffa" (UID: "1bad348b-3d85-4a22-9d03-734715eedffa"). InnerVolumeSpecName "kube-api-access-dbkf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.612787 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bcc82f1-7d2a-48b5-83cb-157f3d4c320c-kube-api-access-wsdxj" (OuterVolumeSpecName: "kube-api-access-wsdxj") pod "7bcc82f1-7d2a-48b5-83cb-157f3d4c320c" (UID: "7bcc82f1-7d2a-48b5-83cb-157f3d4c320c"). InnerVolumeSpecName "kube-api-access-wsdxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.691933 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2540-account-create-w7dp4" event={"ID":"06f78e27-776f-4506-b3f2-a348318cbd13","Type":"ContainerDied","Data":"7951014b50694cb7e4a047e7d5ac204653bb556d9c6caccff9d91d20a4b8677d"} Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.691972 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2540-account-create-w7dp4" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.691980 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7951014b50694cb7e4a047e7d5ac204653bb556d9c6caccff9d91d20a4b8677d" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.693406 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dnl8c" event={"ID":"1bad348b-3d85-4a22-9d03-734715eedffa","Type":"ContainerDied","Data":"42f789eaf4b1ef88016a2f89b119e8e6ee5bd74b7a4748a832dfd653c2bd8e1a"} Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.693423 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42f789eaf4b1ef88016a2f89b119e8e6ee5bd74b7a4748a832dfd653c2bd8e1a" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.693465 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dnl8c" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.694883 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pz5bj" event={"ID":"91db574f-2e79-4e56-a408-64373360aa4a","Type":"ContainerDied","Data":"3f8d085ad57d887407403b371bb7ef758a4dae964924809efd3facff9de4bbfe"} Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.694915 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f8d085ad57d887407403b371bb7ef758a4dae964924809efd3facff9de4bbfe" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.696387 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-00f1-account-create-pfn7s" event={"ID":"7bcc82f1-7d2a-48b5-83cb-157f3d4c320c","Type":"ContainerDied","Data":"5416cb49412a03dab5cdf3aa742a55fa985b37149f235bc66f649522b61f855a"} Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.696407 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5416cb49412a03dab5cdf3aa742a55fa985b37149f235bc66f649522b61f855a" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.696445 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-00f1-account-create-pfn7s" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.698643 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e4c1-account-create-4f5hg" event={"ID":"7c34951d-7bde-4c8e-b369-fc6505c3a752","Type":"ContainerDied","Data":"8d223a28d3c8f3e6f5d0e008f9c1d62d5110812c01a9adf6c3e64c0b7b39250a"} Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.698782 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d223a28d3c8f3e6f5d0e008f9c1d62d5110812c01a9adf6c3e64c0b7b39250a" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.700082 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vxq56" event={"ID":"f4e0697a-3b6b-4d9c-a863-9ef8a696920a","Type":"ContainerDied","Data":"bdc4863e5661c54e551bb8fc869ec484f3d8c028fff26e4c637a1a0914b342c3"} Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.700122 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdc4863e5661c54e551bb8fc869ec484f3d8c028fff26e4c637a1a0914b342c3" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.705125 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pz5bj" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.711940 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06f78e27-776f-4506-b3f2-a348318cbd13-operator-scripts\") pod \"06f78e27-776f-4506-b3f2-a348318cbd13\" (UID: \"06f78e27-776f-4506-b3f2-a348318cbd13\") " Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.716179 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7j586\" (UniqueName: \"kubernetes.io/projected/06f78e27-776f-4506-b3f2-a348318cbd13-kube-api-access-7j586\") pod \"06f78e27-776f-4506-b3f2-a348318cbd13\" (UID: \"06f78e27-776f-4506-b3f2-a348318cbd13\") " Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.712340 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06f78e27-776f-4506-b3f2-a348318cbd13-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "06f78e27-776f-4506-b3f2-a348318cbd13" (UID: "06f78e27-776f-4506-b3f2-a348318cbd13"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.718806 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1bad348b-3d85-4a22-9d03-734715eedffa-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.719263 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06f78e27-776f-4506-b3f2-a348318cbd13-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.719344 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsdxj\" (UniqueName: \"kubernetes.io/projected/7bcc82f1-7d2a-48b5-83cb-157f3d4c320c-kube-api-access-wsdxj\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.719425 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bcc82f1-7d2a-48b5-83cb-157f3d4c320c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.719496 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbkf9\" (UniqueName: \"kubernetes.io/projected/1bad348b-3d85-4a22-9d03-734715eedffa-kube-api-access-dbkf9\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.724354 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06f78e27-776f-4506-b3f2-a348318cbd13-kube-api-access-7j586" (OuterVolumeSpecName: "kube-api-access-7j586") pod "06f78e27-776f-4506-b3f2-a348318cbd13" (UID: "06f78e27-776f-4506-b3f2-a348318cbd13"). InnerVolumeSpecName "kube-api-access-7j586". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.730154 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e4c1-account-create-4f5hg" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.759951 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vxq56" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.820517 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91db574f-2e79-4e56-a408-64373360aa4a-operator-scripts\") pod \"91db574f-2e79-4e56-a408-64373360aa4a\" (UID: \"91db574f-2e79-4e56-a408-64373360aa4a\") " Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.820565 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lf9cv\" (UniqueName: \"kubernetes.io/projected/91db574f-2e79-4e56-a408-64373360aa4a-kube-api-access-lf9cv\") pod \"91db574f-2e79-4e56-a408-64373360aa4a\" (UID: \"91db574f-2e79-4e56-a408-64373360aa4a\") " Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.820688 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5bvj\" (UniqueName: \"kubernetes.io/projected/f4e0697a-3b6b-4d9c-a863-9ef8a696920a-kube-api-access-t5bvj\") pod \"f4e0697a-3b6b-4d9c-a863-9ef8a696920a\" (UID: \"f4e0697a-3b6b-4d9c-a863-9ef8a696920a\") " Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.820937 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91db574f-2e79-4e56-a408-64373360aa4a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "91db574f-2e79-4e56-a408-64373360aa4a" (UID: "91db574f-2e79-4e56-a408-64373360aa4a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.820997 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c34951d-7bde-4c8e-b369-fc6505c3a752-operator-scripts\") pod \"7c34951d-7bde-4c8e-b369-fc6505c3a752\" (UID: \"7c34951d-7bde-4c8e-b369-fc6505c3a752\") " Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.821114 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4e0697a-3b6b-4d9c-a863-9ef8a696920a-operator-scripts\") pod \"f4e0697a-3b6b-4d9c-a863-9ef8a696920a\" (UID: \"f4e0697a-3b6b-4d9c-a863-9ef8a696920a\") " Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.821174 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2np6\" (UniqueName: \"kubernetes.io/projected/7c34951d-7bde-4c8e-b369-fc6505c3a752-kube-api-access-m2np6\") pod \"7c34951d-7bde-4c8e-b369-fc6505c3a752\" (UID: \"7c34951d-7bde-4c8e-b369-fc6505c3a752\") " Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.821522 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c34951d-7bde-4c8e-b369-fc6505c3a752-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7c34951d-7bde-4c8e-b369-fc6505c3a752" (UID: "7c34951d-7bde-4c8e-b369-fc6505c3a752"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.821997 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4e0697a-3b6b-4d9c-a863-9ef8a696920a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f4e0697a-3b6b-4d9c-a863-9ef8a696920a" (UID: "f4e0697a-3b6b-4d9c-a863-9ef8a696920a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.822027 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91db574f-2e79-4e56-a408-64373360aa4a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.822049 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c34951d-7bde-4c8e-b369-fc6505c3a752-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.822062 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7j586\" (UniqueName: \"kubernetes.io/projected/06f78e27-776f-4506-b3f2-a348318cbd13-kube-api-access-7j586\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.824492 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91db574f-2e79-4e56-a408-64373360aa4a-kube-api-access-lf9cv" (OuterVolumeSpecName: "kube-api-access-lf9cv") pod "91db574f-2e79-4e56-a408-64373360aa4a" (UID: "91db574f-2e79-4e56-a408-64373360aa4a"). InnerVolumeSpecName "kube-api-access-lf9cv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.824811 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c34951d-7bde-4c8e-b369-fc6505c3a752-kube-api-access-m2np6" (OuterVolumeSpecName: "kube-api-access-m2np6") pod "7c34951d-7bde-4c8e-b369-fc6505c3a752" (UID: "7c34951d-7bde-4c8e-b369-fc6505c3a752"). InnerVolumeSpecName "kube-api-access-m2np6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.826122 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4e0697a-3b6b-4d9c-a863-9ef8a696920a-kube-api-access-t5bvj" (OuterVolumeSpecName: "kube-api-access-t5bvj") pod "f4e0697a-3b6b-4d9c-a863-9ef8a696920a" (UID: "f4e0697a-3b6b-4d9c-a863-9ef8a696920a"). InnerVolumeSpecName "kube-api-access-t5bvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.923924 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5bvj\" (UniqueName: \"kubernetes.io/projected/f4e0697a-3b6b-4d9c-a863-9ef8a696920a-kube-api-access-t5bvj\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.923983 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4e0697a-3b6b-4d9c-a863-9ef8a696920a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.924002 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2np6\" (UniqueName: \"kubernetes.io/projected/7c34951d-7bde-4c8e-b369-fc6505c3a752-kube-api-access-m2np6\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:29 crc kubenswrapper[4809]: I1124 07:11:29.924015 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lf9cv\" (UniqueName: \"kubernetes.io/projected/91db574f-2e79-4e56-a408-64373360aa4a-kube-api-access-lf9cv\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:30 crc kubenswrapper[4809]: I1124 07:11:30.708384 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vxq56" Nov 24 07:11:30 crc kubenswrapper[4809]: I1124 07:11:30.709317 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-tbhvc" event={"ID":"99c73ccf-06f4-4dfb-ab11-a1b93ab653ea","Type":"ContainerStarted","Data":"05d1220c6bdb4756a0bceb61608c2122633b928859c1d4247efbe42ba4f4d90a"} Nov 24 07:11:30 crc kubenswrapper[4809]: I1124 07:11:30.710377 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e4c1-account-create-4f5hg" Nov 24 07:11:30 crc kubenswrapper[4809]: I1124 07:11:30.712043 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pz5bj" Nov 24 07:11:30 crc kubenswrapper[4809]: I1124 07:11:30.737719 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-tbhvc" podStartSLOduration=2.721274222 podStartE2EDuration="7.737694416s" podCreationTimestamp="2025-11-24 07:11:23 +0000 UTC" firstStartedPulling="2025-11-24 07:11:24.556270295 +0000 UTC m=+984.456862000" lastFinishedPulling="2025-11-24 07:11:29.572690489 +0000 UTC m=+989.473282194" observedRunningTime="2025-11-24 07:11:30.733451464 +0000 UTC m=+990.634043169" watchObservedRunningTime="2025-11-24 07:11:30.737694416 +0000 UTC m=+990.638286121" Nov 24 07:11:32 crc kubenswrapper[4809]: I1124 07:11:32.724714 4809 generic.go:334] "Generic (PLEG): container finished" podID="99c73ccf-06f4-4dfb-ab11-a1b93ab653ea" containerID="05d1220c6bdb4756a0bceb61608c2122633b928859c1d4247efbe42ba4f4d90a" exitCode=0 Nov 24 07:11:32 crc kubenswrapper[4809]: I1124 07:11:32.724797 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-tbhvc" event={"ID":"99c73ccf-06f4-4dfb-ab11-a1b93ab653ea","Type":"ContainerDied","Data":"05d1220c6bdb4756a0bceb61608c2122633b928859c1d4247efbe42ba4f4d90a"} Nov 24 07:11:33 crc kubenswrapper[4809]: I1124 07:11:33.815321 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:11:33 crc kubenswrapper[4809]: I1124 07:11:33.903873 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-jlbv5"] Nov 24 07:11:33 crc kubenswrapper[4809]: I1124 07:11:33.904113 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-jlbv5" podUID="8e6016ff-f68c-423c-bef0-a015064efaaf" containerName="dnsmasq-dns" containerID="cri-o://91115c44f7164e1aac7f11221bb5bd58295543222f37253f77e9c059cb1541ff" gracePeriod=10 Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.156866 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-tbhvc" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.208648 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6nhk\" (UniqueName: \"kubernetes.io/projected/99c73ccf-06f4-4dfb-ab11-a1b93ab653ea-kube-api-access-x6nhk\") pod \"99c73ccf-06f4-4dfb-ab11-a1b93ab653ea\" (UID: \"99c73ccf-06f4-4dfb-ab11-a1b93ab653ea\") " Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.208716 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99c73ccf-06f4-4dfb-ab11-a1b93ab653ea-combined-ca-bundle\") pod \"99c73ccf-06f4-4dfb-ab11-a1b93ab653ea\" (UID: \"99c73ccf-06f4-4dfb-ab11-a1b93ab653ea\") " Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.208819 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99c73ccf-06f4-4dfb-ab11-a1b93ab653ea-config-data\") pod \"99c73ccf-06f4-4dfb-ab11-a1b93ab653ea\" (UID: \"99c73ccf-06f4-4dfb-ab11-a1b93ab653ea\") " Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.221256 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99c73ccf-06f4-4dfb-ab11-a1b93ab653ea-kube-api-access-x6nhk" (OuterVolumeSpecName: "kube-api-access-x6nhk") pod "99c73ccf-06f4-4dfb-ab11-a1b93ab653ea" (UID: "99c73ccf-06f4-4dfb-ab11-a1b93ab653ea"). InnerVolumeSpecName "kube-api-access-x6nhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.236814 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99c73ccf-06f4-4dfb-ab11-a1b93ab653ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99c73ccf-06f4-4dfb-ab11-a1b93ab653ea" (UID: "99c73ccf-06f4-4dfb-ab11-a1b93ab653ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.274400 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99c73ccf-06f4-4dfb-ab11-a1b93ab653ea-config-data" (OuterVolumeSpecName: "config-data") pod "99c73ccf-06f4-4dfb-ab11-a1b93ab653ea" (UID: "99c73ccf-06f4-4dfb-ab11-a1b93ab653ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.311262 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6nhk\" (UniqueName: \"kubernetes.io/projected/99c73ccf-06f4-4dfb-ab11-a1b93ab653ea-kube-api-access-x6nhk\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.311296 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99c73ccf-06f4-4dfb-ab11-a1b93ab653ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.311306 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99c73ccf-06f4-4dfb-ab11-a1b93ab653ea-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.343535 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-jlbv5" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.411868 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-ovsdbserver-nb\") pod \"8e6016ff-f68c-423c-bef0-a015064efaaf\" (UID: \"8e6016ff-f68c-423c-bef0-a015064efaaf\") " Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.412828 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-ovsdbserver-sb\") pod \"8e6016ff-f68c-423c-bef0-a015064efaaf\" (UID: \"8e6016ff-f68c-423c-bef0-a015064efaaf\") " Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.412878 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-config\") pod \"8e6016ff-f68c-423c-bef0-a015064efaaf\" (UID: \"8e6016ff-f68c-423c-bef0-a015064efaaf\") " Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.413031 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrtp8\" (UniqueName: \"kubernetes.io/projected/8e6016ff-f68c-423c-bef0-a015064efaaf-kube-api-access-hrtp8\") pod \"8e6016ff-f68c-423c-bef0-a015064efaaf\" (UID: \"8e6016ff-f68c-423c-bef0-a015064efaaf\") " Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.413056 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-dns-svc\") pod \"8e6016ff-f68c-423c-bef0-a015064efaaf\" (UID: \"8e6016ff-f68c-423c-bef0-a015064efaaf\") " Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.416783 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e6016ff-f68c-423c-bef0-a015064efaaf-kube-api-access-hrtp8" (OuterVolumeSpecName: "kube-api-access-hrtp8") pod "8e6016ff-f68c-423c-bef0-a015064efaaf" (UID: "8e6016ff-f68c-423c-bef0-a015064efaaf"). InnerVolumeSpecName "kube-api-access-hrtp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.456439 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8e6016ff-f68c-423c-bef0-a015064efaaf" (UID: "8e6016ff-f68c-423c-bef0-a015064efaaf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.456620 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8e6016ff-f68c-423c-bef0-a015064efaaf" (UID: "8e6016ff-f68c-423c-bef0-a015064efaaf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.457036 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-config" (OuterVolumeSpecName: "config") pod "8e6016ff-f68c-423c-bef0-a015064efaaf" (UID: "8e6016ff-f68c-423c-bef0-a015064efaaf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.464347 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8e6016ff-f68c-423c-bef0-a015064efaaf" (UID: "8e6016ff-f68c-423c-bef0-a015064efaaf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.514447 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrtp8\" (UniqueName: \"kubernetes.io/projected/8e6016ff-f68c-423c-bef0-a015064efaaf-kube-api-access-hrtp8\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.514480 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.514491 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.514500 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.514507 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e6016ff-f68c-423c-bef0-a015064efaaf-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.743389 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-tbhvc" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.743395 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-tbhvc" event={"ID":"99c73ccf-06f4-4dfb-ab11-a1b93ab653ea","Type":"ContainerDied","Data":"7a83c2085a05d8fdff741cd0d3f85f26c7ed75eae5196fb8599694affbdef865"} Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.743725 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a83c2085a05d8fdff741cd0d3f85f26c7ed75eae5196fb8599694affbdef865" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.745566 4809 generic.go:334] "Generic (PLEG): container finished" podID="8e6016ff-f68c-423c-bef0-a015064efaaf" containerID="91115c44f7164e1aac7f11221bb5bd58295543222f37253f77e9c059cb1541ff" exitCode=0 Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.745608 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-jlbv5" event={"ID":"8e6016ff-f68c-423c-bef0-a015064efaaf","Type":"ContainerDied","Data":"91115c44f7164e1aac7f11221bb5bd58295543222f37253f77e9c059cb1541ff"} Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.745630 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-jlbv5" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.745646 4809 scope.go:117] "RemoveContainer" containerID="91115c44f7164e1aac7f11221bb5bd58295543222f37253f77e9c059cb1541ff" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.745632 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-jlbv5" event={"ID":"8e6016ff-f68c-423c-bef0-a015064efaaf","Type":"ContainerDied","Data":"cc8a404d1bdad2dd6e9ceb8c19ee5d15acd31d58e2990398305d03446eed8ffe"} Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.772405 4809 scope.go:117] "RemoveContainer" containerID="8dbc0a02093b03837764b69c924162211844cb7df74f5c22db0eb1b2eec4d217" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.778805 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-jlbv5"] Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.785068 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-jlbv5"] Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.816828 4809 scope.go:117] "RemoveContainer" containerID="91115c44f7164e1aac7f11221bb5bd58295543222f37253f77e9c059cb1541ff" Nov 24 07:11:34 crc kubenswrapper[4809]: E1124 07:11:34.818417 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91115c44f7164e1aac7f11221bb5bd58295543222f37253f77e9c059cb1541ff\": container with ID starting with 91115c44f7164e1aac7f11221bb5bd58295543222f37253f77e9c059cb1541ff not found: ID does not exist" containerID="91115c44f7164e1aac7f11221bb5bd58295543222f37253f77e9c059cb1541ff" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.818458 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91115c44f7164e1aac7f11221bb5bd58295543222f37253f77e9c059cb1541ff"} err="failed to get container status \"91115c44f7164e1aac7f11221bb5bd58295543222f37253f77e9c059cb1541ff\": rpc error: code = NotFound desc = could not find container \"91115c44f7164e1aac7f11221bb5bd58295543222f37253f77e9c059cb1541ff\": container with ID starting with 91115c44f7164e1aac7f11221bb5bd58295543222f37253f77e9c059cb1541ff not found: ID does not exist" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.818484 4809 scope.go:117] "RemoveContainer" containerID="8dbc0a02093b03837764b69c924162211844cb7df74f5c22db0eb1b2eec4d217" Nov 24 07:11:34 crc kubenswrapper[4809]: E1124 07:11:34.818832 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dbc0a02093b03837764b69c924162211844cb7df74f5c22db0eb1b2eec4d217\": container with ID starting with 8dbc0a02093b03837764b69c924162211844cb7df74f5c22db0eb1b2eec4d217 not found: ID does not exist" containerID="8dbc0a02093b03837764b69c924162211844cb7df74f5c22db0eb1b2eec4d217" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.818869 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dbc0a02093b03837764b69c924162211844cb7df74f5c22db0eb1b2eec4d217"} err="failed to get container status \"8dbc0a02093b03837764b69c924162211844cb7df74f5c22db0eb1b2eec4d217\": rpc error: code = NotFound desc = could not find container \"8dbc0a02093b03837764b69c924162211844cb7df74f5c22db0eb1b2eec4d217\": container with ID starting with 8dbc0a02093b03837764b69c924162211844cb7df74f5c22db0eb1b2eec4d217 not found: ID does not exist" Nov 24 07:11:34 crc kubenswrapper[4809]: I1124 07:11:34.902057 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e6016ff-f68c-423c-bef0-a015064efaaf" path="/var/lib/kubelet/pods/8e6016ff-f68c-423c-bef0-a015064efaaf/volumes" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.417787 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-kmpjx"] Nov 24 07:11:35 crc kubenswrapper[4809]: E1124 07:11:35.418481 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e6016ff-f68c-423c-bef0-a015064efaaf" containerName="dnsmasq-dns" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.418503 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e6016ff-f68c-423c-bef0-a015064efaaf" containerName="dnsmasq-dns" Nov 24 07:11:35 crc kubenswrapper[4809]: E1124 07:11:35.418517 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99c73ccf-06f4-4dfb-ab11-a1b93ab653ea" containerName="keystone-db-sync" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.418525 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="99c73ccf-06f4-4dfb-ab11-a1b93ab653ea" containerName="keystone-db-sync" Nov 24 07:11:35 crc kubenswrapper[4809]: E1124 07:11:35.418538 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9" containerName="dnsmasq-dns" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.418546 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9" containerName="dnsmasq-dns" Nov 24 07:11:35 crc kubenswrapper[4809]: E1124 07:11:35.418560 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c34951d-7bde-4c8e-b369-fc6505c3a752" containerName="mariadb-account-create" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.418567 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c34951d-7bde-4c8e-b369-fc6505c3a752" containerName="mariadb-account-create" Nov 24 07:11:35 crc kubenswrapper[4809]: E1124 07:11:35.418582 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bcc82f1-7d2a-48b5-83cb-157f3d4c320c" containerName="mariadb-account-create" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.418588 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bcc82f1-7d2a-48b5-83cb-157f3d4c320c" containerName="mariadb-account-create" Nov 24 07:11:35 crc kubenswrapper[4809]: E1124 07:11:35.418602 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91db574f-2e79-4e56-a408-64373360aa4a" containerName="mariadb-database-create" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.418609 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="91db574f-2e79-4e56-a408-64373360aa4a" containerName="mariadb-database-create" Nov 24 07:11:35 crc kubenswrapper[4809]: E1124 07:11:35.418624 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e6016ff-f68c-423c-bef0-a015064efaaf" containerName="init" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.418631 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e6016ff-f68c-423c-bef0-a015064efaaf" containerName="init" Nov 24 07:11:35 crc kubenswrapper[4809]: E1124 07:11:35.418649 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9" containerName="init" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.418657 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9" containerName="init" Nov 24 07:11:35 crc kubenswrapper[4809]: E1124 07:11:35.418672 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f78e27-776f-4506-b3f2-a348318cbd13" containerName="mariadb-account-create" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.418679 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f78e27-776f-4506-b3f2-a348318cbd13" containerName="mariadb-account-create" Nov 24 07:11:35 crc kubenswrapper[4809]: E1124 07:11:35.418693 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bad348b-3d85-4a22-9d03-734715eedffa" containerName="mariadb-database-create" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.418700 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bad348b-3d85-4a22-9d03-734715eedffa" containerName="mariadb-database-create" Nov 24 07:11:35 crc kubenswrapper[4809]: E1124 07:11:35.418719 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4e0697a-3b6b-4d9c-a863-9ef8a696920a" containerName="mariadb-database-create" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.418725 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4e0697a-3b6b-4d9c-a863-9ef8a696920a" containerName="mariadb-database-create" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.418883 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4e0697a-3b6b-4d9c-a863-9ef8a696920a" containerName="mariadb-database-create" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.418902 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9de9b687-9a75-41e6-af0e-5dbcf8cbf8a9" containerName="dnsmasq-dns" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.418912 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="91db574f-2e79-4e56-a408-64373360aa4a" containerName="mariadb-database-create" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.418922 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="99c73ccf-06f4-4dfb-ab11-a1b93ab653ea" containerName="keystone-db-sync" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.418936 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e6016ff-f68c-423c-bef0-a015064efaaf" containerName="dnsmasq-dns" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.418952 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bad348b-3d85-4a22-9d03-734715eedffa" containerName="mariadb-database-create" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.418982 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c34951d-7bde-4c8e-b369-fc6505c3a752" containerName="mariadb-account-create" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.418993 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="06f78e27-776f-4506-b3f2-a348318cbd13" containerName="mariadb-account-create" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.419001 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bcc82f1-7d2a-48b5-83cb-157f3d4c320c" containerName="mariadb-account-create" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.419758 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.445073 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-kmpjx"] Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.450851 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-jd86n"] Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.451837 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jd86n" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.453796 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.455236 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.456523 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.456683 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.458568 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fkz7g" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.491318 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jd86n"] Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.613024 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-4l5jw"] Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.621416 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4l5jw" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.629229 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-qgc8p" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.629467 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.632204 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.636560 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-56545965c7-wbhlj"] Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.637266 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx7g9\" (UniqueName: \"kubernetes.io/projected/93c5d091-5b5d-4247-8e56-eefadc0835d1-kube-api-access-xx7g9\") pod \"keystone-bootstrap-jd86n\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " pod="openstack/keystone-bootstrap-jd86n" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.637314 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-kmpjx\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.637344 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-dns-svc\") pod \"dnsmasq-dns-847c4cc679-kmpjx\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.637366 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-credential-keys\") pod \"keystone-bootstrap-jd86n\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " pod="openstack/keystone-bootstrap-jd86n" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.637391 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-config-data\") pod \"keystone-bootstrap-jd86n\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " pod="openstack/keystone-bootstrap-jd86n" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.637430 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrkdp\" (UniqueName: \"kubernetes.io/projected/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-kube-api-access-xrkdp\") pod \"dnsmasq-dns-847c4cc679-kmpjx\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.637447 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-kmpjx\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.637462 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-fernet-keys\") pod \"keystone-bootstrap-jd86n\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " pod="openstack/keystone-bootstrap-jd86n" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.637478 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-kmpjx\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.637514 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-config\") pod \"dnsmasq-dns-847c4cc679-kmpjx\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.637532 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-scripts\") pod \"keystone-bootstrap-jd86n\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " pod="openstack/keystone-bootstrap-jd86n" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.637568 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-combined-ca-bundle\") pod \"keystone-bootstrap-jd86n\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " pod="openstack/keystone-bootstrap-jd86n" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.651426 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-56545965c7-wbhlj" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.666490 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-rbrd7" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.666561 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.666688 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.668366 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.676026 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-dtg82"] Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.677079 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-dtg82" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.681954 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-6m26k" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.682251 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.682403 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.703196 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-4l5jw"] Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.718551 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-56545965c7-wbhlj"] Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.739780 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrkdp\" (UniqueName: \"kubernetes.io/projected/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-kube-api-access-xrkdp\") pod \"dnsmasq-dns-847c4cc679-kmpjx\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.739826 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-kmpjx\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.739841 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-fernet-keys\") pod \"keystone-bootstrap-jd86n\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " pod="openstack/keystone-bootstrap-jd86n" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.739858 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-kmpjx\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.739877 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8cc0b19f-d90f-4460-909e-6eb644cb2955-config\") pod \"neutron-db-sync-4l5jw\" (UID: \"8cc0b19f-d90f-4460-909e-6eb644cb2955\") " pod="openstack/neutron-db-sync-4l5jw" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.739897 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/59e981b3-a120-4a1b-937c-8cc3892392de-scripts\") pod \"horizon-56545965c7-wbhlj\" (UID: \"59e981b3-a120-4a1b-937c-8cc3892392de\") " pod="openstack/horizon-56545965c7-wbhlj" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.739917 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8frz\" (UniqueName: \"kubernetes.io/projected/8cc0b19f-d90f-4460-909e-6eb644cb2955-kube-api-access-t8frz\") pod \"neutron-db-sync-4l5jw\" (UID: \"8cc0b19f-d90f-4460-909e-6eb644cb2955\") " pod="openstack/neutron-db-sync-4l5jw" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.739937 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/59e981b3-a120-4a1b-937c-8cc3892392de-horizon-secret-key\") pod \"horizon-56545965c7-wbhlj\" (UID: \"59e981b3-a120-4a1b-937c-8cc3892392de\") " pod="openstack/horizon-56545965c7-wbhlj" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.739961 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/59e981b3-a120-4a1b-937c-8cc3892392de-config-data\") pod \"horizon-56545965c7-wbhlj\" (UID: \"59e981b3-a120-4a1b-937c-8cc3892392de\") " pod="openstack/horizon-56545965c7-wbhlj" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.739995 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-config\") pod \"dnsmasq-dns-847c4cc679-kmpjx\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.740013 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-scripts\") pod \"keystone-bootstrap-jd86n\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " pod="openstack/keystone-bootstrap-jd86n" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.740034 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cc0b19f-d90f-4460-909e-6eb644cb2955-combined-ca-bundle\") pod \"neutron-db-sync-4l5jw\" (UID: \"8cc0b19f-d90f-4460-909e-6eb644cb2955\") " pod="openstack/neutron-db-sync-4l5jw" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.740058 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59e981b3-a120-4a1b-937c-8cc3892392de-logs\") pod \"horizon-56545965c7-wbhlj\" (UID: \"59e981b3-a120-4a1b-937c-8cc3892392de\") " pod="openstack/horizon-56545965c7-wbhlj" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.740080 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-combined-ca-bundle\") pod \"keystone-bootstrap-jd86n\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " pod="openstack/keystone-bootstrap-jd86n" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.740099 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqslh\" (UniqueName: \"kubernetes.io/projected/59e981b3-a120-4a1b-937c-8cc3892392de-kube-api-access-wqslh\") pod \"horizon-56545965c7-wbhlj\" (UID: \"59e981b3-a120-4a1b-937c-8cc3892392de\") " pod="openstack/horizon-56545965c7-wbhlj" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.740128 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx7g9\" (UniqueName: \"kubernetes.io/projected/93c5d091-5b5d-4247-8e56-eefadc0835d1-kube-api-access-xx7g9\") pod \"keystone-bootstrap-jd86n\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " pod="openstack/keystone-bootstrap-jd86n" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.740149 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-kmpjx\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.740171 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-dns-svc\") pod \"dnsmasq-dns-847c4cc679-kmpjx\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.740189 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-credential-keys\") pod \"keystone-bootstrap-jd86n\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " pod="openstack/keystone-bootstrap-jd86n" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.740212 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-config-data\") pod \"keystone-bootstrap-jd86n\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " pod="openstack/keystone-bootstrap-jd86n" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.749049 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-kmpjx\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.754259 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-config-data\") pod \"keystone-bootstrap-jd86n\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " pod="openstack/keystone-bootstrap-jd86n" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.754266 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-dtg82"] Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.755888 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-fernet-keys\") pod \"keystone-bootstrap-jd86n\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " pod="openstack/keystone-bootstrap-jd86n" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.757085 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-kmpjx\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.757649 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-dns-svc\") pod \"dnsmasq-dns-847c4cc679-kmpjx\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.758305 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-kmpjx\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.765368 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-config\") pod \"dnsmasq-dns-847c4cc679-kmpjx\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.789648 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrkdp\" (UniqueName: \"kubernetes.io/projected/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-kube-api-access-xrkdp\") pod \"dnsmasq-dns-847c4cc679-kmpjx\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.795979 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.804126 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.809862 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.811407 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-scripts\") pod \"keystone-bootstrap-jd86n\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " pod="openstack/keystone-bootstrap-jd86n" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.813584 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-combined-ca-bundle\") pod \"keystone-bootstrap-jd86n\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " pod="openstack/keystone-bootstrap-jd86n" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.832431 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.835040 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx7g9\" (UniqueName: \"kubernetes.io/projected/93c5d091-5b5d-4247-8e56-eefadc0835d1-kube-api-access-xx7g9\") pod \"keystone-bootstrap-jd86n\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " pod="openstack/keystone-bootstrap-jd86n" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.835240 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.835618 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-credential-keys\") pod \"keystone-bootstrap-jd86n\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " pod="openstack/keystone-bootstrap-jd86n" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.841817 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-scripts\") pod \"cinder-db-sync-dtg82\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " pod="openstack/cinder-db-sync-dtg82" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.853626 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9q8n6\" (UniqueName: \"kubernetes.io/projected/35dbbf41-295e-4ec9-bd02-12911c093ce8-kube-api-access-9q8n6\") pod \"cinder-db-sync-dtg82\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " pod="openstack/cinder-db-sync-dtg82" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.854051 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/35dbbf41-295e-4ec9-bd02-12911c093ce8-etc-machine-id\") pod \"cinder-db-sync-dtg82\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " pod="openstack/cinder-db-sync-dtg82" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.854287 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8cc0b19f-d90f-4460-909e-6eb644cb2955-config\") pod \"neutron-db-sync-4l5jw\" (UID: \"8cc0b19f-d90f-4460-909e-6eb644cb2955\") " pod="openstack/neutron-db-sync-4l5jw" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.854368 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-db-sync-config-data\") pod \"cinder-db-sync-dtg82\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " pod="openstack/cinder-db-sync-dtg82" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.854460 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/59e981b3-a120-4a1b-937c-8cc3892392de-scripts\") pod \"horizon-56545965c7-wbhlj\" (UID: \"59e981b3-a120-4a1b-937c-8cc3892392de\") " pod="openstack/horizon-56545965c7-wbhlj" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.854570 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8frz\" (UniqueName: \"kubernetes.io/projected/8cc0b19f-d90f-4460-909e-6eb644cb2955-kube-api-access-t8frz\") pod \"neutron-db-sync-4l5jw\" (UID: \"8cc0b19f-d90f-4460-909e-6eb644cb2955\") " pod="openstack/neutron-db-sync-4l5jw" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.854681 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/59e981b3-a120-4a1b-937c-8cc3892392de-horizon-secret-key\") pod \"horizon-56545965c7-wbhlj\" (UID: \"59e981b3-a120-4a1b-937c-8cc3892392de\") " pod="openstack/horizon-56545965c7-wbhlj" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.854850 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/59e981b3-a120-4a1b-937c-8cc3892392de-config-data\") pod \"horizon-56545965c7-wbhlj\" (UID: \"59e981b3-a120-4a1b-937c-8cc3892392de\") " pod="openstack/horizon-56545965c7-wbhlj" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.854986 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-config-data\") pod \"cinder-db-sync-dtg82\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " pod="openstack/cinder-db-sync-dtg82" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.855075 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-combined-ca-bundle\") pod \"cinder-db-sync-dtg82\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " pod="openstack/cinder-db-sync-dtg82" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.855150 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cc0b19f-d90f-4460-909e-6eb644cb2955-combined-ca-bundle\") pod \"neutron-db-sync-4l5jw\" (UID: \"8cc0b19f-d90f-4460-909e-6eb644cb2955\") " pod="openstack/neutron-db-sync-4l5jw" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.855244 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59e981b3-a120-4a1b-937c-8cc3892392de-logs\") pod \"horizon-56545965c7-wbhlj\" (UID: \"59e981b3-a120-4a1b-937c-8cc3892392de\") " pod="openstack/horizon-56545965c7-wbhlj" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.855345 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqslh\" (UniqueName: \"kubernetes.io/projected/59e981b3-a120-4a1b-937c-8cc3892392de-kube-api-access-wqslh\") pod \"horizon-56545965c7-wbhlj\" (UID: \"59e981b3-a120-4a1b-937c-8cc3892392de\") " pod="openstack/horizon-56545965c7-wbhlj" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.855987 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/59e981b3-a120-4a1b-937c-8cc3892392de-scripts\") pod \"horizon-56545965c7-wbhlj\" (UID: \"59e981b3-a120-4a1b-937c-8cc3892392de\") " pod="openstack/horizon-56545965c7-wbhlj" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.860359 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/59e981b3-a120-4a1b-937c-8cc3892392de-config-data\") pod \"horizon-56545965c7-wbhlj\" (UID: \"59e981b3-a120-4a1b-937c-8cc3892392de\") " pod="openstack/horizon-56545965c7-wbhlj" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.860599 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59e981b3-a120-4a1b-937c-8cc3892392de-logs\") pod \"horizon-56545965c7-wbhlj\" (UID: \"59e981b3-a120-4a1b-937c-8cc3892392de\") " pod="openstack/horizon-56545965c7-wbhlj" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.871992 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8cc0b19f-d90f-4460-909e-6eb644cb2955-config\") pod \"neutron-db-sync-4l5jw\" (UID: \"8cc0b19f-d90f-4460-909e-6eb644cb2955\") " pod="openstack/neutron-db-sync-4l5jw" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.872755 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/59e981b3-a120-4a1b-937c-8cc3892392de-horizon-secret-key\") pod \"horizon-56545965c7-wbhlj\" (UID: \"59e981b3-a120-4a1b-937c-8cc3892392de\") " pod="openstack/horizon-56545965c7-wbhlj" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.879929 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cc0b19f-d90f-4460-909e-6eb644cb2955-combined-ca-bundle\") pod \"neutron-db-sync-4l5jw\" (UID: \"8cc0b19f-d90f-4460-909e-6eb644cb2955\") " pod="openstack/neutron-db-sync-4l5jw" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.895086 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-kmpjx"] Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.895604 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.920849 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqslh\" (UniqueName: \"kubernetes.io/projected/59e981b3-a120-4a1b-937c-8cc3892392de-kube-api-access-wqslh\") pod \"horizon-56545965c7-wbhlj\" (UID: \"59e981b3-a120-4a1b-937c-8cc3892392de\") " pod="openstack/horizon-56545965c7-wbhlj" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.926909 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.928477 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.933744 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8frz\" (UniqueName: \"kubernetes.io/projected/8cc0b19f-d90f-4460-909e-6eb644cb2955-kube-api-access-t8frz\") pod \"neutron-db-sync-4l5jw\" (UID: \"8cc0b19f-d90f-4460-909e-6eb644cb2955\") " pod="openstack/neutron-db-sync-4l5jw" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.941005 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.941203 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.941318 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-cwg4f" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.942910 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.950005 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-lkj6z"] Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.956902 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " pod="openstack/ceilometer-0" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.957026 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " pod="openstack/ceilometer-0" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.957067 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-scripts\") pod \"cinder-db-sync-dtg82\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " pod="openstack/cinder-db-sync-dtg82" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.957088 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-config-data\") pod \"ceilometer-0\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " pod="openstack/ceilometer-0" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.957110 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9q8n6\" (UniqueName: \"kubernetes.io/projected/35dbbf41-295e-4ec9-bd02-12911c093ce8-kube-api-access-9q8n6\") pod \"cinder-db-sync-dtg82\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " pod="openstack/cinder-db-sync-dtg82" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.957127 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhprc\" (UniqueName: \"kubernetes.io/projected/655e6590-673e-4fde-a097-1173dfcef9e2-kube-api-access-mhprc\") pod \"ceilometer-0\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " pod="openstack/ceilometer-0" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.957171 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/35dbbf41-295e-4ec9-bd02-12911c093ce8-etc-machine-id\") pod \"cinder-db-sync-dtg82\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " pod="openstack/cinder-db-sync-dtg82" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.957204 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-scripts\") pod \"ceilometer-0\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " pod="openstack/ceilometer-0" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.957226 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-db-sync-config-data\") pod \"cinder-db-sync-dtg82\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " pod="openstack/cinder-db-sync-dtg82" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.957247 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/655e6590-673e-4fde-a097-1173dfcef9e2-log-httpd\") pod \"ceilometer-0\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " pod="openstack/ceilometer-0" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.957268 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/655e6590-673e-4fde-a097-1173dfcef9e2-run-httpd\") pod \"ceilometer-0\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " pod="openstack/ceilometer-0" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.957305 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-config-data\") pod \"cinder-db-sync-dtg82\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " pod="openstack/cinder-db-sync-dtg82" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.957319 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-combined-ca-bundle\") pod \"cinder-db-sync-dtg82\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " pod="openstack/cinder-db-sync-dtg82" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.962263 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-lkj6z" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.963281 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/35dbbf41-295e-4ec9-bd02-12911c093ce8-etc-machine-id\") pod \"cinder-db-sync-dtg82\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " pod="openstack/cinder-db-sync-dtg82" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.964353 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.964553 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-db-sync-config-data\") pod \"cinder-db-sync-dtg82\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " pod="openstack/cinder-db-sync-dtg82" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.964597 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-dvvzk" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.964834 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4l5jw" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.964916 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-scripts\") pod \"cinder-db-sync-dtg82\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " pod="openstack/cinder-db-sync-dtg82" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.966409 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-combined-ca-bundle\") pod \"cinder-db-sync-dtg82\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " pod="openstack/cinder-db-sync-dtg82" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.978980 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-56545965c7-wbhlj" Nov 24 07:11:35 crc kubenswrapper[4809]: I1124 07:11:35.986904 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-config-data\") pod \"cinder-db-sync-dtg82\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " pod="openstack/cinder-db-sync-dtg82" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.013715 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9q8n6\" (UniqueName: \"kubernetes.io/projected/35dbbf41-295e-4ec9-bd02-12911c093ce8-kube-api-access-9q8n6\") pod \"cinder-db-sync-dtg82\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " pod="openstack/cinder-db-sync-dtg82" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.063102 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-546d49d8b5-77xgm"] Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.064432 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-546d49d8b5-77xgm" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.065828 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.065917 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-scripts\") pod \"ceilometer-0\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " pod="openstack/ceilometer-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.065948 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2z62x\" (UniqueName: \"kubernetes.io/projected/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-kube-api-access-2z62x\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.065983 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/655e6590-673e-4fde-a097-1173dfcef9e2-log-httpd\") pod \"ceilometer-0\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " pod="openstack/ceilometer-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.066003 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/655e6590-673e-4fde-a097-1173dfcef9e2-run-httpd\") pod \"ceilometer-0\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " pod="openstack/ceilometer-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.066020 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-scripts\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.066047 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.066074 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-logs\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.066097 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.066117 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " pod="openstack/ceilometer-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.066133 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf79z\" (UniqueName: \"kubernetes.io/projected/972cb0ce-ea08-408c-a10d-ba1d14d42568-kube-api-access-lf79z\") pod \"barbican-db-sync-lkj6z\" (UID: \"972cb0ce-ea08-408c-a10d-ba1d14d42568\") " pod="openstack/barbican-db-sync-lkj6z" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.066152 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/972cb0ce-ea08-408c-a10d-ba1d14d42568-db-sync-config-data\") pod \"barbican-db-sync-lkj6z\" (UID: \"972cb0ce-ea08-408c-a10d-ba1d14d42568\") " pod="openstack/barbican-db-sync-lkj6z" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.066167 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/972cb0ce-ea08-408c-a10d-ba1d14d42568-combined-ca-bundle\") pod \"barbican-db-sync-lkj6z\" (UID: \"972cb0ce-ea08-408c-a10d-ba1d14d42568\") " pod="openstack/barbican-db-sync-lkj6z" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.066185 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-config-data\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.066206 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " pod="openstack/ceilometer-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.066225 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.066257 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-config-data\") pod \"ceilometer-0\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " pod="openstack/ceilometer-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.066310 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhprc\" (UniqueName: \"kubernetes.io/projected/655e6590-673e-4fde-a097-1173dfcef9e2-kube-api-access-mhprc\") pod \"ceilometer-0\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " pod="openstack/ceilometer-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.072856 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/655e6590-673e-4fde-a097-1173dfcef9e2-log-httpd\") pod \"ceilometer-0\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " pod="openstack/ceilometer-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.073127 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/655e6590-673e-4fde-a097-1173dfcef9e2-run-httpd\") pod \"ceilometer-0\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " pod="openstack/ceilometer-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.077351 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-lkj6z"] Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.084328 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jd86n" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.097416 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " pod="openstack/ceilometer-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.100154 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " pod="openstack/ceilometer-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.100594 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-scripts\") pod \"ceilometer-0\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " pod="openstack/ceilometer-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.101224 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.104799 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-config-data\") pod \"ceilometer-0\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " pod="openstack/ceilometer-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.110047 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhprc\" (UniqueName: \"kubernetes.io/projected/655e6590-673e-4fde-a097-1173dfcef9e2-kube-api-access-mhprc\") pod \"ceilometer-0\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " pod="openstack/ceilometer-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.111052 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-ncjd6"] Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.112044 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ncjd6" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.115355 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.115619 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.115943 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-5xnm8" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.118230 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-ncjd6"] Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.127109 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-546d49d8b5-77xgm"] Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.130669 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-pvgsp"] Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.132166 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.143516 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-pvgsp"] Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.156629 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.158107 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.160622 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.160917 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.166936 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.168322 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2z62x\" (UniqueName: \"kubernetes.io/projected/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-kube-api-access-2z62x\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.168374 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e04f0b8-285c-4539-9996-9df4429c7bf8-config-data\") pod \"horizon-546d49d8b5-77xgm\" (UID: \"9e04f0b8-285c-4539-9996-9df4429c7bf8\") " pod="openstack/horizon-546d49d8b5-77xgm" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.168404 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-scripts\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.168435 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.168459 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-logs\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.168497 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.168527 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf79z\" (UniqueName: \"kubernetes.io/projected/972cb0ce-ea08-408c-a10d-ba1d14d42568-kube-api-access-lf79z\") pod \"barbican-db-sync-lkj6z\" (UID: \"972cb0ce-ea08-408c-a10d-ba1d14d42568\") " pod="openstack/barbican-db-sync-lkj6z" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.168554 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/972cb0ce-ea08-408c-a10d-ba1d14d42568-combined-ca-bundle\") pod \"barbican-db-sync-lkj6z\" (UID: \"972cb0ce-ea08-408c-a10d-ba1d14d42568\") " pod="openstack/barbican-db-sync-lkj6z" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.168577 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/972cb0ce-ea08-408c-a10d-ba1d14d42568-db-sync-config-data\") pod \"barbican-db-sync-lkj6z\" (UID: \"972cb0ce-ea08-408c-a10d-ba1d14d42568\") " pod="openstack/barbican-db-sync-lkj6z" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.168604 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e04f0b8-285c-4539-9996-9df4429c7bf8-scripts\") pod \"horizon-546d49d8b5-77xgm\" (UID: \"9e04f0b8-285c-4539-9996-9df4429c7bf8\") " pod="openstack/horizon-546d49d8b5-77xgm" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.168630 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-config-data\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.168661 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.168709 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e04f0b8-285c-4539-9996-9df4429c7bf8-logs\") pod \"horizon-546d49d8b5-77xgm\" (UID: \"9e04f0b8-285c-4539-9996-9df4429c7bf8\") " pod="openstack/horizon-546d49d8b5-77xgm" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.168734 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmtwm\" (UniqueName: \"kubernetes.io/projected/9e04f0b8-285c-4539-9996-9df4429c7bf8-kube-api-access-kmtwm\") pod \"horizon-546d49d8b5-77xgm\" (UID: \"9e04f0b8-285c-4539-9996-9df4429c7bf8\") " pod="openstack/horizon-546d49d8b5-77xgm" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.168770 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9e04f0b8-285c-4539-9996-9df4429c7bf8-horizon-secret-key\") pod \"horizon-546d49d8b5-77xgm\" (UID: \"9e04f0b8-285c-4539-9996-9df4429c7bf8\") " pod="openstack/horizon-546d49d8b5-77xgm" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.168828 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.170753 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.187610 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.188217 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.188688 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/972cb0ce-ea08-408c-a10d-ba1d14d42568-db-sync-config-data\") pod \"barbican-db-sync-lkj6z\" (UID: \"972cb0ce-ea08-408c-a10d-ba1d14d42568\") " pod="openstack/barbican-db-sync-lkj6z" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.190856 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-logs\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.242129 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/972cb0ce-ea08-408c-a10d-ba1d14d42568-combined-ca-bundle\") pod \"barbican-db-sync-lkj6z\" (UID: \"972cb0ce-ea08-408c-a10d-ba1d14d42568\") " pod="openstack/barbican-db-sync-lkj6z" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.268710 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-config-data\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.273707 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-scripts\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.275267 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-pvgsp\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.275307 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.275629 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-pvgsp\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.275688 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-combined-ca-bundle\") pod \"placement-db-sync-ncjd6\" (UID: \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\") " pod="openstack/placement-db-sync-ncjd6" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.275720 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-logs\") pod \"placement-db-sync-ncjd6\" (UID: \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\") " pod="openstack/placement-db-sync-ncjd6" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.275734 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.275760 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e04f0b8-285c-4539-9996-9df4429c7bf8-scripts\") pod \"horizon-546d49d8b5-77xgm\" (UID: \"9e04f0b8-285c-4539-9996-9df4429c7bf8\") " pod="openstack/horizon-546d49d8b5-77xgm" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.275780 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.275804 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-logs\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.275830 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-pvgsp\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.275855 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e04f0b8-285c-4539-9996-9df4429c7bf8-logs\") pod \"horizon-546d49d8b5-77xgm\" (UID: \"9e04f0b8-285c-4539-9996-9df4429c7bf8\") " pod="openstack/horizon-546d49d8b5-77xgm" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.275869 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmtwm\" (UniqueName: \"kubernetes.io/projected/9e04f0b8-285c-4539-9996-9df4429c7bf8-kube-api-access-kmtwm\") pod \"horizon-546d49d8b5-77xgm\" (UID: \"9e04f0b8-285c-4539-9996-9df4429c7bf8\") " pod="openstack/horizon-546d49d8b5-77xgm" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.275889 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.275908 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9e04f0b8-285c-4539-9996-9df4429c7bf8-horizon-secret-key\") pod \"horizon-546d49d8b5-77xgm\" (UID: \"9e04f0b8-285c-4539-9996-9df4429c7bf8\") " pod="openstack/horizon-546d49d8b5-77xgm" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.275936 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.275977 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6sz4\" (UniqueName: \"kubernetes.io/projected/946235c6-45ea-4bfe-98c1-24a8095b1843-kube-api-access-h6sz4\") pod \"dnsmasq-dns-785d8bcb8c-pvgsp\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.275994 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-config-data\") pod \"placement-db-sync-ncjd6\" (UID: \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\") " pod="openstack/placement-db-sync-ncjd6" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.276012 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.276033 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-scripts\") pod \"placement-db-sync-ncjd6\" (UID: \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\") " pod="openstack/placement-db-sync-ncjd6" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.276069 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-pvgsp\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.276098 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e04f0b8-285c-4539-9996-9df4429c7bf8-config-data\") pod \"horizon-546d49d8b5-77xgm\" (UID: \"9e04f0b8-285c-4539-9996-9df4429c7bf8\") " pod="openstack/horizon-546d49d8b5-77xgm" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.276986 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e04f0b8-285c-4539-9996-9df4429c7bf8-scripts\") pod \"horizon-546d49d8b5-77xgm\" (UID: \"9e04f0b8-285c-4539-9996-9df4429c7bf8\") " pod="openstack/horizon-546d49d8b5-77xgm" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.277269 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e04f0b8-285c-4539-9996-9df4429c7bf8-logs\") pod \"horizon-546d49d8b5-77xgm\" (UID: \"9e04f0b8-285c-4539-9996-9df4429c7bf8\") " pod="openstack/horizon-546d49d8b5-77xgm" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.280832 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.281902 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e04f0b8-285c-4539-9996-9df4429c7bf8-config-data\") pod \"horizon-546d49d8b5-77xgm\" (UID: \"9e04f0b8-285c-4539-9996-9df4429c7bf8\") " pod="openstack/horizon-546d49d8b5-77xgm" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.282808 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2z62x\" (UniqueName: \"kubernetes.io/projected/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-kube-api-access-2z62x\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.287653 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf79z\" (UniqueName: \"kubernetes.io/projected/972cb0ce-ea08-408c-a10d-ba1d14d42568-kube-api-access-lf79z\") pod \"barbican-db-sync-lkj6z\" (UID: \"972cb0ce-ea08-408c-a10d-ba1d14d42568\") " pod="openstack/barbican-db-sync-lkj6z" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.288196 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.294427 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9e04f0b8-285c-4539-9996-9df4429c7bf8-horizon-secret-key\") pod \"horizon-546d49d8b5-77xgm\" (UID: \"9e04f0b8-285c-4539-9996-9df4429c7bf8\") " pod="openstack/horizon-546d49d8b5-77xgm" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.300717 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7bg7\" (UniqueName: \"kubernetes.io/projected/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-kube-api-access-t7bg7\") pod \"placement-db-sync-ncjd6\" (UID: \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\") " pod="openstack/placement-db-sync-ncjd6" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.300743 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbpnz\" (UniqueName: \"kubernetes.io/projected/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-kube-api-access-pbpnz\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.300764 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-config\") pod \"dnsmasq-dns-785d8bcb8c-pvgsp\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.310230 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-dtg82" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.334586 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmtwm\" (UniqueName: \"kubernetes.io/projected/9e04f0b8-285c-4539-9996-9df4429c7bf8-kube-api-access-kmtwm\") pod \"horizon-546d49d8b5-77xgm\" (UID: \"9e04f0b8-285c-4539-9996-9df4429c7bf8\") " pod="openstack/horizon-546d49d8b5-77xgm" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.351604 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-lkj6z" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.363395 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-546d49d8b5-77xgm" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.364469 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.401860 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.401910 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.401938 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6sz4\" (UniqueName: \"kubernetes.io/projected/946235c6-45ea-4bfe-98c1-24a8095b1843-kube-api-access-h6sz4\") pod \"dnsmasq-dns-785d8bcb8c-pvgsp\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.401958 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-config-data\") pod \"placement-db-sync-ncjd6\" (UID: \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\") " pod="openstack/placement-db-sync-ncjd6" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.401990 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.402010 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-scripts\") pod \"placement-db-sync-ncjd6\" (UID: \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\") " pod="openstack/placement-db-sync-ncjd6" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.402042 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-pvgsp\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.402063 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7bg7\" (UniqueName: \"kubernetes.io/projected/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-kube-api-access-t7bg7\") pod \"placement-db-sync-ncjd6\" (UID: \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\") " pod="openstack/placement-db-sync-ncjd6" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.402080 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbpnz\" (UniqueName: \"kubernetes.io/projected/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-kube-api-access-pbpnz\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.402097 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-config\") pod \"dnsmasq-dns-785d8bcb8c-pvgsp\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.402122 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-pvgsp\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.402140 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.402163 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-pvgsp\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.402184 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-combined-ca-bundle\") pod \"placement-db-sync-ncjd6\" (UID: \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\") " pod="openstack/placement-db-sync-ncjd6" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.402206 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-logs\") pod \"placement-db-sync-ncjd6\" (UID: \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\") " pod="openstack/placement-db-sync-ncjd6" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.402222 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.402240 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.402262 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-logs\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.402288 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-pvgsp\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.403038 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-pvgsp\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.406471 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-pvgsp\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.409574 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-pvgsp\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.409995 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.410151 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-logs\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.410186 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.410449 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-config\") pod \"dnsmasq-dns-785d8bcb8c-pvgsp\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.410727 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.410756 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.410784 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.412532 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.443588 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-config-data\") pod \"placement-db-sync-ncjd6\" (UID: \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\") " pod="openstack/placement-db-sync-ncjd6" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.444664 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-pvgsp\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.446905 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-logs\") pod \"placement-db-sync-ncjd6\" (UID: \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\") " pod="openstack/placement-db-sync-ncjd6" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.457017 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-scripts\") pod \"placement-db-sync-ncjd6\" (UID: \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\") " pod="openstack/placement-db-sync-ncjd6" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.458773 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6sz4\" (UniqueName: \"kubernetes.io/projected/946235c6-45ea-4bfe-98c1-24a8095b1843-kube-api-access-h6sz4\") pod \"dnsmasq-dns-785d8bcb8c-pvgsp\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.462594 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7bg7\" (UniqueName: \"kubernetes.io/projected/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-kube-api-access-t7bg7\") pod \"placement-db-sync-ncjd6\" (UID: \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\") " pod="openstack/placement-db-sync-ncjd6" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.463958 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbpnz\" (UniqueName: \"kubernetes.io/projected/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-kube-api-access-pbpnz\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.489923 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-combined-ca-bundle\") pod \"placement-db-sync-ncjd6\" (UID: \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\") " pod="openstack/placement-db-sync-ncjd6" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.500439 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.631354 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.686847 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ncjd6" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.698027 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.706196 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.709650 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-4l5jw"] Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.771433 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-kmpjx"] Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.806592 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4l5jw" event={"ID":"8cc0b19f-d90f-4460-909e-6eb644cb2955","Type":"ContainerStarted","Data":"7144dc03fe2711c0989bb80d008b55405d3366cfffa8d2ee9a89ff62a35d8da8"} Nov 24 07:11:36 crc kubenswrapper[4809]: I1124 07:11:36.831294 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-56545965c7-wbhlj"] Nov 24 07:11:36 crc kubenswrapper[4809]: W1124 07:11:36.838794 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5f9e7b3_2680_48f1_a85e_f14b9b7d9f33.slice/crio-9c225b111d6f972028e8948d79cc7eac0b5e08200e5aa70b009680b2cd515826 WatchSource:0}: Error finding container 9c225b111d6f972028e8948d79cc7eac0b5e08200e5aa70b009680b2cd515826: Status 404 returned error can't find the container with id 9c225b111d6f972028e8948d79cc7eac0b5e08200e5aa70b009680b2cd515826 Nov 24 07:11:36 crc kubenswrapper[4809]: W1124 07:11:36.859922 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59e981b3_a120_4a1b_937c_8cc3892392de.slice/crio-92f4ed3d551452ef2b1e4c77f0f367d7059974d41f63b70cdbbcd81b18d95a13 WatchSource:0}: Error finding container 92f4ed3d551452ef2b1e4c77f0f367d7059974d41f63b70cdbbcd81b18d95a13: Status 404 returned error can't find the container with id 92f4ed3d551452ef2b1e4c77f0f367d7059974d41f63b70cdbbcd81b18d95a13 Nov 24 07:11:37 crc kubenswrapper[4809]: I1124 07:11:37.042212 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jd86n"] Nov 24 07:11:37 crc kubenswrapper[4809]: I1124 07:11:37.161048 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:11:37 crc kubenswrapper[4809]: I1124 07:11:37.834988 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jd86n" event={"ID":"93c5d091-5b5d-4247-8e56-eefadc0835d1","Type":"ContainerStarted","Data":"b70526d14ae2999033922d90886b26dd0c6962c84bdd995e51968f2ff939118c"} Nov 24 07:11:37 crc kubenswrapper[4809]: I1124 07:11:37.835029 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jd86n" event={"ID":"93c5d091-5b5d-4247-8e56-eefadc0835d1","Type":"ContainerStarted","Data":"aeeb46187be88e328d197e5032a0ed8d6a9ef1ec4f2f99160c3758e4dc327eb4"} Nov 24 07:11:37 crc kubenswrapper[4809]: I1124 07:11:37.841421 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4l5jw" event={"ID":"8cc0b19f-d90f-4460-909e-6eb644cb2955","Type":"ContainerStarted","Data":"2c1da7380ece890327fe29583a5580de0895e39d1b91fc06834a63134f70c22a"} Nov 24 07:11:37 crc kubenswrapper[4809]: I1124 07:11:37.856145 4809 generic.go:334] "Generic (PLEG): container finished" podID="d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33" containerID="b316a4344eb01d40eacdfcbc4d31b2823f4f541c320577b7097471655ceb71f2" exitCode=0 Nov 24 07:11:37 crc kubenswrapper[4809]: I1124 07:11:37.856222 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" event={"ID":"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33","Type":"ContainerDied","Data":"b316a4344eb01d40eacdfcbc4d31b2823f4f541c320577b7097471655ceb71f2"} Nov 24 07:11:37 crc kubenswrapper[4809]: I1124 07:11:37.856251 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" event={"ID":"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33","Type":"ContainerStarted","Data":"9c225b111d6f972028e8948d79cc7eac0b5e08200e5aa70b009680b2cd515826"} Nov 24 07:11:37 crc kubenswrapper[4809]: I1124 07:11:37.861472 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-jd86n" podStartSLOduration=2.86145368 podStartE2EDuration="2.86145368s" podCreationTimestamp="2025-11-24 07:11:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:11:37.852292107 +0000 UTC m=+997.752883812" watchObservedRunningTime="2025-11-24 07:11:37.86145368 +0000 UTC m=+997.762045385" Nov 24 07:11:37 crc kubenswrapper[4809]: I1124 07:11:37.867437 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-56545965c7-wbhlj" event={"ID":"59e981b3-a120-4a1b-937c-8cc3892392de","Type":"ContainerStarted","Data":"92f4ed3d551452ef2b1e4c77f0f367d7059974d41f63b70cdbbcd81b18d95a13"} Nov 24 07:11:37 crc kubenswrapper[4809]: I1124 07:11:37.869165 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"655e6590-673e-4fde-a097-1173dfcef9e2","Type":"ContainerStarted","Data":"b80b1b435358522aa4a781ad61588386f6857b29c7ea99d9f87106d3726e9d0e"} Nov 24 07:11:37 crc kubenswrapper[4809]: I1124 07:11:37.876388 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-4l5jw" podStartSLOduration=2.876369835 podStartE2EDuration="2.876369835s" podCreationTimestamp="2025-11-24 07:11:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:11:37.865702843 +0000 UTC m=+997.766294548" watchObservedRunningTime="2025-11-24 07:11:37.876369835 +0000 UTC m=+997.776961540" Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.074251 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-dtg82"] Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.090113 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-546d49d8b5-77xgm"] Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.103716 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-lkj6z"] Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.406308 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-ncjd6"] Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.419139 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-pvgsp"] Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.518546 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.521204 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.564946 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-ovsdbserver-sb\") pod \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.565035 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrkdp\" (UniqueName: \"kubernetes.io/projected/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-kube-api-access-xrkdp\") pod \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.565090 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-config\") pod \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.565141 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-ovsdbserver-nb\") pod \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.565196 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-dns-svc\") pod \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.565228 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-dns-swift-storage-0\") pod \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\" (UID: \"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33\") " Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.587206 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-kube-api-access-xrkdp" (OuterVolumeSpecName: "kube-api-access-xrkdp") pod "d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33" (UID: "d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33"). InnerVolumeSpecName "kube-api-access-xrkdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.588204 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33" (UID: "d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.592398 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33" (UID: "d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.594483 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-config" (OuterVolumeSpecName: "config") pod "d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33" (UID: "d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.597074 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33" (UID: "d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.619577 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33" (UID: "d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.666991 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.667042 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrkdp\" (UniqueName: \"kubernetes.io/projected/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-kube-api-access-xrkdp\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.667054 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.667062 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.667070 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.667079 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.889814 4809 generic.go:334] "Generic (PLEG): container finished" podID="946235c6-45ea-4bfe-98c1-24a8095b1843" containerID="53b48d5ec8d1e3104e2c4a8ee666ec65ede0579a328ca3d848707ff42f8bb98a" exitCode=0 Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.889880 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" event={"ID":"946235c6-45ea-4bfe-98c1-24a8095b1843","Type":"ContainerDied","Data":"53b48d5ec8d1e3104e2c4a8ee666ec65ede0579a328ca3d848707ff42f8bb98a"} Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.889905 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" event={"ID":"946235c6-45ea-4bfe-98c1-24a8095b1843","Type":"ContainerStarted","Data":"b085f99ad2c9e6b59d590a563ebf6cd6582ed3f3fe42f1a6558bd628afe4fc13"} Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.896782 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.923668 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-dtg82" event={"ID":"35dbbf41-295e-4ec9-bd02-12911c093ce8","Type":"ContainerStarted","Data":"54e9341ce2cb545e234b9294c9745fb3e37ad1a25f0074ca172c45b53ec66ef3"} Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.923710 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4","Type":"ContainerStarted","Data":"459a35f6f64d83ee8ee13446cf9e3f7a6f1adf176576b930c284de24887afb8d"} Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.923730 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-kmpjx" event={"ID":"d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33","Type":"ContainerDied","Data":"9c225b111d6f972028e8948d79cc7eac0b5e08200e5aa70b009680b2cd515826"} Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.923746 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-546d49d8b5-77xgm" event={"ID":"9e04f0b8-285c-4539-9996-9df4429c7bf8","Type":"ContainerStarted","Data":"dd3f12cb214167bc81f8dede51e2a991e120821ebc16b856e61021e3b0936c51"} Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.923754 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ncjd6" event={"ID":"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293","Type":"ContainerStarted","Data":"205905a1eddb1ff10ccefeaf24de603b33d2aeab2b41f9058b5c7991d4c231b9"} Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.923763 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-lkj6z" event={"ID":"972cb0ce-ea08-408c-a10d-ba1d14d42568","Type":"ContainerStarted","Data":"d8c83e0e63e7f83fcb57551ac76044b91d9fd7031f91729f158ae5ba92506e3e"} Nov 24 07:11:38 crc kubenswrapper[4809]: I1124 07:11:38.923778 4809 scope.go:117] "RemoveContainer" containerID="b316a4344eb01d40eacdfcbc4d31b2823f4f541c320577b7097471655ceb71f2" Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.016420 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.044547 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-kmpjx"] Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.059177 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-kmpjx"] Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.065668 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-546d49d8b5-77xgm"] Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.088328 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-84666554f7-jc9pp"] Nov 24 07:11:39 crc kubenswrapper[4809]: E1124 07:11:39.088768 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33" containerName="init" Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.088796 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33" containerName="init" Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.089010 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33" containerName="init" Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.089936 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-84666554f7-jc9pp" Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.112874 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-84666554f7-jc9pp"] Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.121541 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.131178 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.183404 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46ebff1a-add8-420f-99c2-c593d486d2f8-logs\") pod \"horizon-84666554f7-jc9pp\" (UID: \"46ebff1a-add8-420f-99c2-c593d486d2f8\") " pod="openstack/horizon-84666554f7-jc9pp" Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.183436 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/46ebff1a-add8-420f-99c2-c593d486d2f8-config-data\") pod \"horizon-84666554f7-jc9pp\" (UID: \"46ebff1a-add8-420f-99c2-c593d486d2f8\") " pod="openstack/horizon-84666554f7-jc9pp" Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.183516 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/46ebff1a-add8-420f-99c2-c593d486d2f8-horizon-secret-key\") pod \"horizon-84666554f7-jc9pp\" (UID: \"46ebff1a-add8-420f-99c2-c593d486d2f8\") " pod="openstack/horizon-84666554f7-jc9pp" Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.183540 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42s9t\" (UniqueName: \"kubernetes.io/projected/46ebff1a-add8-420f-99c2-c593d486d2f8-kube-api-access-42s9t\") pod \"horizon-84666554f7-jc9pp\" (UID: \"46ebff1a-add8-420f-99c2-c593d486d2f8\") " pod="openstack/horizon-84666554f7-jc9pp" Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.183556 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46ebff1a-add8-420f-99c2-c593d486d2f8-scripts\") pod \"horizon-84666554f7-jc9pp\" (UID: \"46ebff1a-add8-420f-99c2-c593d486d2f8\") " pod="openstack/horizon-84666554f7-jc9pp" Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.284577 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46ebff1a-add8-420f-99c2-c593d486d2f8-logs\") pod \"horizon-84666554f7-jc9pp\" (UID: \"46ebff1a-add8-420f-99c2-c593d486d2f8\") " pod="openstack/horizon-84666554f7-jc9pp" Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.284621 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/46ebff1a-add8-420f-99c2-c593d486d2f8-config-data\") pod \"horizon-84666554f7-jc9pp\" (UID: \"46ebff1a-add8-420f-99c2-c593d486d2f8\") " pod="openstack/horizon-84666554f7-jc9pp" Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.286004 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/46ebff1a-add8-420f-99c2-c593d486d2f8-config-data\") pod \"horizon-84666554f7-jc9pp\" (UID: \"46ebff1a-add8-420f-99c2-c593d486d2f8\") " pod="openstack/horizon-84666554f7-jc9pp" Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.286081 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/46ebff1a-add8-420f-99c2-c593d486d2f8-horizon-secret-key\") pod \"horizon-84666554f7-jc9pp\" (UID: \"46ebff1a-add8-420f-99c2-c593d486d2f8\") " pod="openstack/horizon-84666554f7-jc9pp" Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.286118 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42s9t\" (UniqueName: \"kubernetes.io/projected/46ebff1a-add8-420f-99c2-c593d486d2f8-kube-api-access-42s9t\") pod \"horizon-84666554f7-jc9pp\" (UID: \"46ebff1a-add8-420f-99c2-c593d486d2f8\") " pod="openstack/horizon-84666554f7-jc9pp" Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.286136 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46ebff1a-add8-420f-99c2-c593d486d2f8-scripts\") pod \"horizon-84666554f7-jc9pp\" (UID: \"46ebff1a-add8-420f-99c2-c593d486d2f8\") " pod="openstack/horizon-84666554f7-jc9pp" Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.286676 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46ebff1a-add8-420f-99c2-c593d486d2f8-scripts\") pod \"horizon-84666554f7-jc9pp\" (UID: \"46ebff1a-add8-420f-99c2-c593d486d2f8\") " pod="openstack/horizon-84666554f7-jc9pp" Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.287247 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46ebff1a-add8-420f-99c2-c593d486d2f8-logs\") pod \"horizon-84666554f7-jc9pp\" (UID: \"46ebff1a-add8-420f-99c2-c593d486d2f8\") " pod="openstack/horizon-84666554f7-jc9pp" Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.295613 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/46ebff1a-add8-420f-99c2-c593d486d2f8-horizon-secret-key\") pod \"horizon-84666554f7-jc9pp\" (UID: \"46ebff1a-add8-420f-99c2-c593d486d2f8\") " pod="openstack/horizon-84666554f7-jc9pp" Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.331239 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42s9t\" (UniqueName: \"kubernetes.io/projected/46ebff1a-add8-420f-99c2-c593d486d2f8-kube-api-access-42s9t\") pod \"horizon-84666554f7-jc9pp\" (UID: \"46ebff1a-add8-420f-99c2-c593d486d2f8\") " pod="openstack/horizon-84666554f7-jc9pp" Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.365158 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.484034 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-84666554f7-jc9pp" Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.961824 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4","Type":"ContainerStarted","Data":"b233c6ea9c0134a2310d53352dd483c95217c6794362c9c5dc9b25069e495e3e"} Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.966093 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1","Type":"ContainerStarted","Data":"07f98d4a21a2357312e0bc1588f4b7c6346845b0b0ad09c05a2c2d0c7b5eac0f"} Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.969081 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" event={"ID":"946235c6-45ea-4bfe-98c1-24a8095b1843","Type":"ContainerStarted","Data":"cb2b20004eca9a344d43c7fccb0cdf95cbe2ef16a181f66d0db8c1eaeca59b56"} Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.971095 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:11:39 crc kubenswrapper[4809]: I1124 07:11:39.989822 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" podStartSLOduration=4.9898033569999996 podStartE2EDuration="4.989803357s" podCreationTimestamp="2025-11-24 07:11:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:11:39.986257323 +0000 UTC m=+999.886849028" watchObservedRunningTime="2025-11-24 07:11:39.989803357 +0000 UTC m=+999.890395062" Nov 24 07:11:40 crc kubenswrapper[4809]: I1124 07:11:40.039905 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-84666554f7-jc9pp"] Nov 24 07:11:40 crc kubenswrapper[4809]: W1124 07:11:40.051883 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46ebff1a_add8_420f_99c2_c593d486d2f8.slice/crio-a8d53e6b56d3b2ddefb6923c5f754fe54a0862ed69f12be9f539be0d5558a4a3 WatchSource:0}: Error finding container a8d53e6b56d3b2ddefb6923c5f754fe54a0862ed69f12be9f539be0d5558a4a3: Status 404 returned error can't find the container with id a8d53e6b56d3b2ddefb6923c5f754fe54a0862ed69f12be9f539be0d5558a4a3 Nov 24 07:11:40 crc kubenswrapper[4809]: I1124 07:11:40.912795 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33" path="/var/lib/kubelet/pods/d5f9e7b3-2680-48f1-a85e-f14b9b7d9f33/volumes" Nov 24 07:11:41 crc kubenswrapper[4809]: I1124 07:11:41.002225 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1","Type":"ContainerStarted","Data":"eeca6acca22917a6233a98d4cb48281b10499512c51d94c995a6831ab1332f5f"} Nov 24 07:11:41 crc kubenswrapper[4809]: I1124 07:11:41.004849 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84666554f7-jc9pp" event={"ID":"46ebff1a-add8-420f-99c2-c593d486d2f8","Type":"ContainerStarted","Data":"a8d53e6b56d3b2ddefb6923c5f754fe54a0862ed69f12be9f539be0d5558a4a3"} Nov 24 07:11:41 crc kubenswrapper[4809]: I1124 07:11:41.010663 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4","Type":"ContainerStarted","Data":"07b8047b10191667e2554f1262e98edd9ac90565c43aa23e76a0a199917caa80"} Nov 24 07:11:41 crc kubenswrapper[4809]: I1124 07:11:41.011047 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="07a92f93-a9f5-4c97-a903-2ab6d43a2ae4" containerName="glance-log" containerID="cri-o://b233c6ea9c0134a2310d53352dd483c95217c6794362c9c5dc9b25069e495e3e" gracePeriod=30 Nov 24 07:11:41 crc kubenswrapper[4809]: I1124 07:11:41.011271 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="07a92f93-a9f5-4c97-a903-2ab6d43a2ae4" containerName="glance-httpd" containerID="cri-o://07b8047b10191667e2554f1262e98edd9ac90565c43aa23e76a0a199917caa80" gracePeriod=30 Nov 24 07:11:41 crc kubenswrapper[4809]: I1124 07:11:41.121122 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.12109964 podStartE2EDuration="6.12109964s" podCreationTimestamp="2025-11-24 07:11:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:11:41.117305539 +0000 UTC m=+1001.017897244" watchObservedRunningTime="2025-11-24 07:11:41.12109964 +0000 UTC m=+1001.021691345" Nov 24 07:11:42 crc kubenswrapper[4809]: I1124 07:11:42.029700 4809 generic.go:334] "Generic (PLEG): container finished" podID="07a92f93-a9f5-4c97-a903-2ab6d43a2ae4" containerID="07b8047b10191667e2554f1262e98edd9ac90565c43aa23e76a0a199917caa80" exitCode=0 Nov 24 07:11:42 crc kubenswrapper[4809]: I1124 07:11:42.029992 4809 generic.go:334] "Generic (PLEG): container finished" podID="07a92f93-a9f5-4c97-a903-2ab6d43a2ae4" containerID="b233c6ea9c0134a2310d53352dd483c95217c6794362c9c5dc9b25069e495e3e" exitCode=143 Nov 24 07:11:42 crc kubenswrapper[4809]: I1124 07:11:42.029755 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4","Type":"ContainerDied","Data":"07b8047b10191667e2554f1262e98edd9ac90565c43aa23e76a0a199917caa80"} Nov 24 07:11:42 crc kubenswrapper[4809]: I1124 07:11:42.030105 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4","Type":"ContainerDied","Data":"b233c6ea9c0134a2310d53352dd483c95217c6794362c9c5dc9b25069e495e3e"} Nov 24 07:11:42 crc kubenswrapper[4809]: I1124 07:11:42.032873 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1","Type":"ContainerStarted","Data":"d8e155e95784d8f39829d39a50c7dc5d5f880fca4903019a1b55780069d2ef57"} Nov 24 07:11:42 crc kubenswrapper[4809]: I1124 07:11:42.033389 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1" containerName="glance-log" containerID="cri-o://eeca6acca22917a6233a98d4cb48281b10499512c51d94c995a6831ab1332f5f" gracePeriod=30 Nov 24 07:11:42 crc kubenswrapper[4809]: I1124 07:11:42.033435 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1" containerName="glance-httpd" containerID="cri-o://d8e155e95784d8f39829d39a50c7dc5d5f880fca4903019a1b55780069d2ef57" gracePeriod=30 Nov 24 07:11:42 crc kubenswrapper[4809]: I1124 07:11:42.051839 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.051818825 podStartE2EDuration="6.051818825s" podCreationTimestamp="2025-11-24 07:11:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:11:42.051545738 +0000 UTC m=+1001.952137453" watchObservedRunningTime="2025-11-24 07:11:42.051818825 +0000 UTC m=+1001.952410530" Nov 24 07:11:43 crc kubenswrapper[4809]: I1124 07:11:43.043546 4809 generic.go:334] "Generic (PLEG): container finished" podID="b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1" containerID="d8e155e95784d8f39829d39a50c7dc5d5f880fca4903019a1b55780069d2ef57" exitCode=0 Nov 24 07:11:43 crc kubenswrapper[4809]: I1124 07:11:43.043920 4809 generic.go:334] "Generic (PLEG): container finished" podID="b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1" containerID="eeca6acca22917a6233a98d4cb48281b10499512c51d94c995a6831ab1332f5f" exitCode=143 Nov 24 07:11:43 crc kubenswrapper[4809]: I1124 07:11:43.043634 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1","Type":"ContainerDied","Data":"d8e155e95784d8f39829d39a50c7dc5d5f880fca4903019a1b55780069d2ef57"} Nov 24 07:11:43 crc kubenswrapper[4809]: I1124 07:11:43.044059 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1","Type":"ContainerDied","Data":"eeca6acca22917a6233a98d4cb48281b10499512c51d94c995a6831ab1332f5f"} Nov 24 07:11:43 crc kubenswrapper[4809]: I1124 07:11:43.045416 4809 generic.go:334] "Generic (PLEG): container finished" podID="93c5d091-5b5d-4247-8e56-eefadc0835d1" containerID="b70526d14ae2999033922d90886b26dd0c6962c84bdd995e51968f2ff939118c" exitCode=0 Nov 24 07:11:43 crc kubenswrapper[4809]: I1124 07:11:43.045452 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jd86n" event={"ID":"93c5d091-5b5d-4247-8e56-eefadc0835d1","Type":"ContainerDied","Data":"b70526d14ae2999033922d90886b26dd0c6962c84bdd995e51968f2ff939118c"} Nov 24 07:11:44 crc kubenswrapper[4809]: I1124 07:11:44.788927 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-56545965c7-wbhlj"] Nov 24 07:11:44 crc kubenswrapper[4809]: I1124 07:11:44.817091 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-79cbd445d4-5hkc8"] Nov 24 07:11:44 crc kubenswrapper[4809]: I1124 07:11:44.831483 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:11:44 crc kubenswrapper[4809]: I1124 07:11:44.836108 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Nov 24 07:11:44 crc kubenswrapper[4809]: I1124 07:11:44.839925 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79cbd445d4-5hkc8"] Nov 24 07:11:44 crc kubenswrapper[4809]: I1124 07:11:44.908990 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b3e7ca6-2814-4a14-b1f4-765f8086927e-logs\") pod \"horizon-79cbd445d4-5hkc8\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:11:44 crc kubenswrapper[4809]: I1124 07:11:44.909317 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5b3e7ca6-2814-4a14-b1f4-765f8086927e-horizon-secret-key\") pod \"horizon-79cbd445d4-5hkc8\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:11:44 crc kubenswrapper[4809]: I1124 07:11:44.909354 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5b3e7ca6-2814-4a14-b1f4-765f8086927e-scripts\") pod \"horizon-79cbd445d4-5hkc8\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:11:44 crc kubenswrapper[4809]: I1124 07:11:44.909410 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b3e7ca6-2814-4a14-b1f4-765f8086927e-combined-ca-bundle\") pod \"horizon-79cbd445d4-5hkc8\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:11:44 crc kubenswrapper[4809]: I1124 07:11:44.909426 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b3e7ca6-2814-4a14-b1f4-765f8086927e-horizon-tls-certs\") pod \"horizon-79cbd445d4-5hkc8\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:11:44 crc kubenswrapper[4809]: I1124 07:11:44.909451 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbv7j\" (UniqueName: \"kubernetes.io/projected/5b3e7ca6-2814-4a14-b1f4-765f8086927e-kube-api-access-jbv7j\") pod \"horizon-79cbd445d4-5hkc8\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:11:44 crc kubenswrapper[4809]: I1124 07:11:44.909492 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5b3e7ca6-2814-4a14-b1f4-765f8086927e-config-data\") pod \"horizon-79cbd445d4-5hkc8\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:11:44 crc kubenswrapper[4809]: I1124 07:11:44.935024 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-84666554f7-jc9pp"] Nov 24 07:11:44 crc kubenswrapper[4809]: I1124 07:11:44.966853 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-8585547888-qdlwv"] Nov 24 07:11:44 crc kubenswrapper[4809]: I1124 07:11:44.969147 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:11:44 crc kubenswrapper[4809]: I1124 07:11:44.979012 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8585547888-qdlwv"] Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.012939 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5b3e7ca6-2814-4a14-b1f4-765f8086927e-horizon-secret-key\") pod \"horizon-79cbd445d4-5hkc8\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.013082 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b5487494-236a-4335-977f-11167eaa3cdc-horizon-secret-key\") pod \"horizon-8585547888-qdlwv\" (UID: \"b5487494-236a-4335-977f-11167eaa3cdc\") " pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.013111 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5487494-236a-4335-977f-11167eaa3cdc-combined-ca-bundle\") pod \"horizon-8585547888-qdlwv\" (UID: \"b5487494-236a-4335-977f-11167eaa3cdc\") " pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.013256 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5b3e7ca6-2814-4a14-b1f4-765f8086927e-scripts\") pod \"horizon-79cbd445d4-5hkc8\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.013377 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5487494-236a-4335-977f-11167eaa3cdc-horizon-tls-certs\") pod \"horizon-8585547888-qdlwv\" (UID: \"b5487494-236a-4335-977f-11167eaa3cdc\") " pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.013511 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhj7r\" (UniqueName: \"kubernetes.io/projected/b5487494-236a-4335-977f-11167eaa3cdc-kube-api-access-jhj7r\") pod \"horizon-8585547888-qdlwv\" (UID: \"b5487494-236a-4335-977f-11167eaa3cdc\") " pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.013547 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b3e7ca6-2814-4a14-b1f4-765f8086927e-combined-ca-bundle\") pod \"horizon-79cbd445d4-5hkc8\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.013637 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b3e7ca6-2814-4a14-b1f4-765f8086927e-horizon-tls-certs\") pod \"horizon-79cbd445d4-5hkc8\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.014516 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5b3e7ca6-2814-4a14-b1f4-765f8086927e-scripts\") pod \"horizon-79cbd445d4-5hkc8\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.015606 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbv7j\" (UniqueName: \"kubernetes.io/projected/5b3e7ca6-2814-4a14-b1f4-765f8086927e-kube-api-access-jbv7j\") pod \"horizon-79cbd445d4-5hkc8\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.015705 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5b3e7ca6-2814-4a14-b1f4-765f8086927e-config-data\") pod \"horizon-79cbd445d4-5hkc8\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.015743 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b5487494-236a-4335-977f-11167eaa3cdc-config-data\") pod \"horizon-8585547888-qdlwv\" (UID: \"b5487494-236a-4335-977f-11167eaa3cdc\") " pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.015804 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b3e7ca6-2814-4a14-b1f4-765f8086927e-logs\") pod \"horizon-79cbd445d4-5hkc8\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.015926 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b5487494-236a-4335-977f-11167eaa3cdc-scripts\") pod \"horizon-8585547888-qdlwv\" (UID: \"b5487494-236a-4335-977f-11167eaa3cdc\") " pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.016025 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5487494-236a-4335-977f-11167eaa3cdc-logs\") pod \"horizon-8585547888-qdlwv\" (UID: \"b5487494-236a-4335-977f-11167eaa3cdc\") " pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.016454 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b3e7ca6-2814-4a14-b1f4-765f8086927e-logs\") pod \"horizon-79cbd445d4-5hkc8\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.020534 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5b3e7ca6-2814-4a14-b1f4-765f8086927e-horizon-secret-key\") pod \"horizon-79cbd445d4-5hkc8\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.020677 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b3e7ca6-2814-4a14-b1f4-765f8086927e-combined-ca-bundle\") pod \"horizon-79cbd445d4-5hkc8\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.021454 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5b3e7ca6-2814-4a14-b1f4-765f8086927e-config-data\") pod \"horizon-79cbd445d4-5hkc8\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.037858 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b3e7ca6-2814-4a14-b1f4-765f8086927e-horizon-tls-certs\") pod \"horizon-79cbd445d4-5hkc8\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.038567 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbv7j\" (UniqueName: \"kubernetes.io/projected/5b3e7ca6-2814-4a14-b1f4-765f8086927e-kube-api-access-jbv7j\") pod \"horizon-79cbd445d4-5hkc8\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.118036 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b5487494-236a-4335-977f-11167eaa3cdc-horizon-secret-key\") pod \"horizon-8585547888-qdlwv\" (UID: \"b5487494-236a-4335-977f-11167eaa3cdc\") " pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.118091 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5487494-236a-4335-977f-11167eaa3cdc-combined-ca-bundle\") pod \"horizon-8585547888-qdlwv\" (UID: \"b5487494-236a-4335-977f-11167eaa3cdc\") " pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.118151 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5487494-236a-4335-977f-11167eaa3cdc-horizon-tls-certs\") pod \"horizon-8585547888-qdlwv\" (UID: \"b5487494-236a-4335-977f-11167eaa3cdc\") " pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.118171 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhj7r\" (UniqueName: \"kubernetes.io/projected/b5487494-236a-4335-977f-11167eaa3cdc-kube-api-access-jhj7r\") pod \"horizon-8585547888-qdlwv\" (UID: \"b5487494-236a-4335-977f-11167eaa3cdc\") " pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.118228 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b5487494-236a-4335-977f-11167eaa3cdc-config-data\") pod \"horizon-8585547888-qdlwv\" (UID: \"b5487494-236a-4335-977f-11167eaa3cdc\") " pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.118294 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b5487494-236a-4335-977f-11167eaa3cdc-scripts\") pod \"horizon-8585547888-qdlwv\" (UID: \"b5487494-236a-4335-977f-11167eaa3cdc\") " pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.118320 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5487494-236a-4335-977f-11167eaa3cdc-logs\") pod \"horizon-8585547888-qdlwv\" (UID: \"b5487494-236a-4335-977f-11167eaa3cdc\") " pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.118795 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5487494-236a-4335-977f-11167eaa3cdc-logs\") pod \"horizon-8585547888-qdlwv\" (UID: \"b5487494-236a-4335-977f-11167eaa3cdc\") " pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.121118 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b5487494-236a-4335-977f-11167eaa3cdc-config-data\") pod \"horizon-8585547888-qdlwv\" (UID: \"b5487494-236a-4335-977f-11167eaa3cdc\") " pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.123510 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5487494-236a-4335-977f-11167eaa3cdc-combined-ca-bundle\") pod \"horizon-8585547888-qdlwv\" (UID: \"b5487494-236a-4335-977f-11167eaa3cdc\") " pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.124597 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b5487494-236a-4335-977f-11167eaa3cdc-scripts\") pod \"horizon-8585547888-qdlwv\" (UID: \"b5487494-236a-4335-977f-11167eaa3cdc\") " pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.127379 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b5487494-236a-4335-977f-11167eaa3cdc-horizon-secret-key\") pod \"horizon-8585547888-qdlwv\" (UID: \"b5487494-236a-4335-977f-11167eaa3cdc\") " pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.143463 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhj7r\" (UniqueName: \"kubernetes.io/projected/b5487494-236a-4335-977f-11167eaa3cdc-kube-api-access-jhj7r\") pod \"horizon-8585547888-qdlwv\" (UID: \"b5487494-236a-4335-977f-11167eaa3cdc\") " pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.145940 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5487494-236a-4335-977f-11167eaa3cdc-horizon-tls-certs\") pod \"horizon-8585547888-qdlwv\" (UID: \"b5487494-236a-4335-977f-11167eaa3cdc\") " pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.160572 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:11:45 crc kubenswrapper[4809]: I1124 07:11:45.298034 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:11:46 crc kubenswrapper[4809]: I1124 07:11:46.700742 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:11:46 crc kubenswrapper[4809]: I1124 07:11:46.773511 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-2b7fh"] Nov 24 07:11:46 crc kubenswrapper[4809]: I1124 07:11:46.773734 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" podUID="1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5" containerName="dnsmasq-dns" containerID="cri-o://95db14eaac6f10051b603ed6695d3a7360907b8e405a445d744910d120259bbe" gracePeriod=10 Nov 24 07:11:47 crc kubenswrapper[4809]: I1124 07:11:47.099592 4809 generic.go:334] "Generic (PLEG): container finished" podID="1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5" containerID="95db14eaac6f10051b603ed6695d3a7360907b8e405a445d744910d120259bbe" exitCode=0 Nov 24 07:11:47 crc kubenswrapper[4809]: I1124 07:11:47.099636 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" event={"ID":"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5","Type":"ContainerDied","Data":"95db14eaac6f10051b603ed6695d3a7360907b8e405a445d744910d120259bbe"} Nov 24 07:11:48 crc kubenswrapper[4809]: I1124 07:11:48.814455 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" podUID="1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.131:5353: connect: connection refused" Nov 24 07:11:51 crc kubenswrapper[4809]: I1124 07:11:51.473933 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jd86n" Nov 24 07:11:51 crc kubenswrapper[4809]: I1124 07:11:51.554810 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-fernet-keys\") pod \"93c5d091-5b5d-4247-8e56-eefadc0835d1\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " Nov 24 07:11:51 crc kubenswrapper[4809]: I1124 07:11:51.555168 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-credential-keys\") pod \"93c5d091-5b5d-4247-8e56-eefadc0835d1\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " Nov 24 07:11:51 crc kubenswrapper[4809]: I1124 07:11:51.555213 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-combined-ca-bundle\") pod \"93c5d091-5b5d-4247-8e56-eefadc0835d1\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " Nov 24 07:11:51 crc kubenswrapper[4809]: I1124 07:11:51.555246 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-scripts\") pod \"93c5d091-5b5d-4247-8e56-eefadc0835d1\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " Nov 24 07:11:51 crc kubenswrapper[4809]: I1124 07:11:51.555381 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-config-data\") pod \"93c5d091-5b5d-4247-8e56-eefadc0835d1\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " Nov 24 07:11:51 crc kubenswrapper[4809]: I1124 07:11:51.555453 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xx7g9\" (UniqueName: \"kubernetes.io/projected/93c5d091-5b5d-4247-8e56-eefadc0835d1-kube-api-access-xx7g9\") pod \"93c5d091-5b5d-4247-8e56-eefadc0835d1\" (UID: \"93c5d091-5b5d-4247-8e56-eefadc0835d1\") " Nov 24 07:11:51 crc kubenswrapper[4809]: I1124 07:11:51.562488 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "93c5d091-5b5d-4247-8e56-eefadc0835d1" (UID: "93c5d091-5b5d-4247-8e56-eefadc0835d1"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:11:51 crc kubenswrapper[4809]: I1124 07:11:51.563284 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93c5d091-5b5d-4247-8e56-eefadc0835d1-kube-api-access-xx7g9" (OuterVolumeSpecName: "kube-api-access-xx7g9") pod "93c5d091-5b5d-4247-8e56-eefadc0835d1" (UID: "93c5d091-5b5d-4247-8e56-eefadc0835d1"). InnerVolumeSpecName "kube-api-access-xx7g9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:11:51 crc kubenswrapper[4809]: I1124 07:11:51.564145 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-scripts" (OuterVolumeSpecName: "scripts") pod "93c5d091-5b5d-4247-8e56-eefadc0835d1" (UID: "93c5d091-5b5d-4247-8e56-eefadc0835d1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:11:51 crc kubenswrapper[4809]: I1124 07:11:51.566189 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "93c5d091-5b5d-4247-8e56-eefadc0835d1" (UID: "93c5d091-5b5d-4247-8e56-eefadc0835d1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:11:51 crc kubenswrapper[4809]: I1124 07:11:51.599874 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "93c5d091-5b5d-4247-8e56-eefadc0835d1" (UID: "93c5d091-5b5d-4247-8e56-eefadc0835d1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:11:51 crc kubenswrapper[4809]: I1124 07:11:51.618006 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-config-data" (OuterVolumeSpecName: "config-data") pod "93c5d091-5b5d-4247-8e56-eefadc0835d1" (UID: "93c5d091-5b5d-4247-8e56-eefadc0835d1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:11:51 crc kubenswrapper[4809]: I1124 07:11:51.659335 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:51 crc kubenswrapper[4809]: I1124 07:11:51.659380 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xx7g9\" (UniqueName: \"kubernetes.io/projected/93c5d091-5b5d-4247-8e56-eefadc0835d1-kube-api-access-xx7g9\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:51 crc kubenswrapper[4809]: I1124 07:11:51.659394 4809 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:51 crc kubenswrapper[4809]: I1124 07:11:51.659408 4809 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:51 crc kubenswrapper[4809]: I1124 07:11:51.659420 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:51 crc kubenswrapper[4809]: I1124 07:11:51.659429 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93c5d091-5b5d-4247-8e56-eefadc0835d1-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.170776 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jd86n" event={"ID":"93c5d091-5b5d-4247-8e56-eefadc0835d1","Type":"ContainerDied","Data":"aeeb46187be88e328d197e5032a0ed8d6a9ef1ec4f2f99160c3758e4dc327eb4"} Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.170821 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aeeb46187be88e328d197e5032a0ed8d6a9ef1ec4f2f99160c3758e4dc327eb4" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.170872 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jd86n" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.580664 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-jd86n"] Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.589496 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-jd86n"] Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.705044 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-qsvl5"] Nov 24 07:11:52 crc kubenswrapper[4809]: E1124 07:11:52.705473 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93c5d091-5b5d-4247-8e56-eefadc0835d1" containerName="keystone-bootstrap" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.705492 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="93c5d091-5b5d-4247-8e56-eefadc0835d1" containerName="keystone-bootstrap" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.705710 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="93c5d091-5b5d-4247-8e56-eefadc0835d1" containerName="keystone-bootstrap" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.706507 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qsvl5" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.709848 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.709891 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.710172 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.710684 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.712557 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fkz7g" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.731323 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-qsvl5"] Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.780505 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-scripts\") pod \"keystone-bootstrap-qsvl5\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " pod="openstack/keystone-bootstrap-qsvl5" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.780655 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-combined-ca-bundle\") pod \"keystone-bootstrap-qsvl5\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " pod="openstack/keystone-bootstrap-qsvl5" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.780689 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w9df\" (UniqueName: \"kubernetes.io/projected/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-kube-api-access-4w9df\") pod \"keystone-bootstrap-qsvl5\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " pod="openstack/keystone-bootstrap-qsvl5" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.780719 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-credential-keys\") pod \"keystone-bootstrap-qsvl5\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " pod="openstack/keystone-bootstrap-qsvl5" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.780786 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-config-data\") pod \"keystone-bootstrap-qsvl5\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " pod="openstack/keystone-bootstrap-qsvl5" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.780920 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-fernet-keys\") pod \"keystone-bootstrap-qsvl5\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " pod="openstack/keystone-bootstrap-qsvl5" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.882077 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-scripts\") pod \"keystone-bootstrap-qsvl5\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " pod="openstack/keystone-bootstrap-qsvl5" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.882192 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-combined-ca-bundle\") pod \"keystone-bootstrap-qsvl5\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " pod="openstack/keystone-bootstrap-qsvl5" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.882219 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w9df\" (UniqueName: \"kubernetes.io/projected/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-kube-api-access-4w9df\") pod \"keystone-bootstrap-qsvl5\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " pod="openstack/keystone-bootstrap-qsvl5" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.882249 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-credential-keys\") pod \"keystone-bootstrap-qsvl5\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " pod="openstack/keystone-bootstrap-qsvl5" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.882279 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-config-data\") pod \"keystone-bootstrap-qsvl5\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " pod="openstack/keystone-bootstrap-qsvl5" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.882339 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-fernet-keys\") pod \"keystone-bootstrap-qsvl5\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " pod="openstack/keystone-bootstrap-qsvl5" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.887248 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-config-data\") pod \"keystone-bootstrap-qsvl5\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " pod="openstack/keystone-bootstrap-qsvl5" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.887408 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-credential-keys\") pod \"keystone-bootstrap-qsvl5\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " pod="openstack/keystone-bootstrap-qsvl5" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.887408 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-fernet-keys\") pod \"keystone-bootstrap-qsvl5\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " pod="openstack/keystone-bootstrap-qsvl5" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.893323 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-scripts\") pod \"keystone-bootstrap-qsvl5\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " pod="openstack/keystone-bootstrap-qsvl5" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.894949 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-combined-ca-bundle\") pod \"keystone-bootstrap-qsvl5\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " pod="openstack/keystone-bootstrap-qsvl5" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.900751 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93c5d091-5b5d-4247-8e56-eefadc0835d1" path="/var/lib/kubelet/pods/93c5d091-5b5d-4247-8e56-eefadc0835d1/volumes" Nov 24 07:11:52 crc kubenswrapper[4809]: I1124 07:11:52.905000 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w9df\" (UniqueName: \"kubernetes.io/projected/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-kube-api-access-4w9df\") pod \"keystone-bootstrap-qsvl5\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " pod="openstack/keystone-bootstrap-qsvl5" Nov 24 07:11:53 crc kubenswrapper[4809]: I1124 07:11:53.021948 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qsvl5" Nov 24 07:11:55 crc kubenswrapper[4809]: E1124 07:11:55.723861 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Nov 24 07:11:55 crc kubenswrapper[4809]: E1124 07:11:55.724303 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t7bg7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-ncjd6_openstack(6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 07:11:55 crc kubenswrapper[4809]: E1124 07:11:55.725499 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-ncjd6" podUID="6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293" Nov 24 07:11:55 crc kubenswrapper[4809]: E1124 07:11:55.752492 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Nov 24 07:11:55 crc kubenswrapper[4809]: E1124 07:11:55.752898 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5f4h5dch685h77hd9h658hb8h5b5h5c9h674hd4h5b6h59h64bh55fh5bh5f7h667h5cch589h57h679h8ch57fh7dh66bh66fh647h88hb7hbbh5c7q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wqslh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-56545965c7-wbhlj_openstack(59e981b3-a120-4a1b-937c-8cc3892392de): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 07:11:55 crc kubenswrapper[4809]: E1124 07:11:55.755931 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-56545965c7-wbhlj" podUID="59e981b3-a120-4a1b-937c-8cc3892392de" Nov 24 07:11:55 crc kubenswrapper[4809]: E1124 07:11:55.769721 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Nov 24 07:11:55 crc kubenswrapper[4809]: E1124 07:11:55.769916 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5cfh646h596h558h54h5b5h594h5h578h55dh8fh5f7h5d4h656h55fh5dfh9ch586h588hcch599h554h544h684h66bh8fh699h679h8dh585h5dch56cq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kmtwm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-546d49d8b5-77xgm_openstack(9e04f0b8-285c-4539-9996-9df4429c7bf8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 07:11:55 crc kubenswrapper[4809]: E1124 07:11:55.772216 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-546d49d8b5-77xgm" podUID="9e04f0b8-285c-4539-9996-9df4429c7bf8" Nov 24 07:11:56 crc kubenswrapper[4809]: E1124 07:11:56.215026 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-ncjd6" podUID="6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293" Nov 24 07:11:58 crc kubenswrapper[4809]: I1124 07:11:58.814714 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" podUID="1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.131:5353: i/o timeout" Nov 24 07:12:03 crc kubenswrapper[4809]: I1124 07:12:03.816272 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" podUID="1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.131:5353: i/o timeout" Nov 24 07:12:03 crc kubenswrapper[4809]: I1124 07:12:03.816910 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:12:03 crc kubenswrapper[4809]: I1124 07:12:03.912764 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.052610 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-httpd-run\") pod \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.052996 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.053160 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-scripts\") pod \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.053283 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-config-data\") pod \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.053417 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-logs\") pod \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.053528 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-internal-tls-certs\") pod \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.053637 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbpnz\" (UniqueName: \"kubernetes.io/projected/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-kube-api-access-pbpnz\") pod \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.053753 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-combined-ca-bundle\") pod \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\" (UID: \"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.053547 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1" (UID: "b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.053650 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-logs" (OuterVolumeSpecName: "logs") pod "b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1" (UID: "b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.054555 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.054839 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.059164 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-kube-api-access-pbpnz" (OuterVolumeSpecName: "kube-api-access-pbpnz") pod "b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1" (UID: "b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1"). InnerVolumeSpecName "kube-api-access-pbpnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.060223 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1" (UID: "b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.073516 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-scripts" (OuterVolumeSpecName: "scripts") pod "b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1" (UID: "b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.085759 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1" (UID: "b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.101232 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1" (UID: "b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.121329 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-config-data" (OuterVolumeSpecName: "config-data") pod "b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1" (UID: "b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.156648 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.156683 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.156693 4809 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.156704 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbpnz\" (UniqueName: \"kubernetes.io/projected/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-kube-api-access-pbpnz\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.156713 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.156742 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.180210 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.258230 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.290652 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1","Type":"ContainerDied","Data":"07f98d4a21a2357312e0bc1588f4b7c6346845b0b0ad09c05a2c2d0c7b5eac0f"} Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.290692 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.290714 4809 scope.go:117] "RemoveContainer" containerID="d8e155e95784d8f39829d39a50c7dc5d5f880fca4903019a1b55780069d2ef57" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.294306 4809 generic.go:334] "Generic (PLEG): container finished" podID="8cc0b19f-d90f-4460-909e-6eb644cb2955" containerID="2c1da7380ece890327fe29583a5580de0895e39d1b91fc06834a63134f70c22a" exitCode=0 Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.294369 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4l5jw" event={"ID":"8cc0b19f-d90f-4460-909e-6eb644cb2955","Type":"ContainerDied","Data":"2c1da7380ece890327fe29583a5580de0895e39d1b91fc06834a63134f70c22a"} Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.331848 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.343661 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.348597 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:12:04 crc kubenswrapper[4809]: E1124 07:12:04.349074 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1" containerName="glance-log" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.349114 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1" containerName="glance-log" Nov 24 07:12:04 crc kubenswrapper[4809]: E1124 07:12:04.349126 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1" containerName="glance-httpd" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.349133 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1" containerName="glance-httpd" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.349307 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1" containerName="glance-httpd" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.349331 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1" containerName="glance-log" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.350270 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.353346 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.353510 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.356424 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.408064 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.423459 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-56545965c7-wbhlj" Nov 24 07:12:04 crc kubenswrapper[4809]: E1124 07:12:04.429534 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Nov 24 07:12:04 crc kubenswrapper[4809]: E1124 07:12:04.429688 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lf79z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-lkj6z_openstack(972cb0ce-ea08-408c-a10d-ba1d14d42568): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 07:12:04 crc kubenswrapper[4809]: E1124 07:12:04.430950 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-lkj6z" podUID="972cb0ce-ea08-408c-a10d-ba1d14d42568" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.448244 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-546d49d8b5-77xgm" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.451133 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.472227 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-logs\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.472372 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.472397 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.472469 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.472487 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.472544 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.472597 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.472718 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl578\" (UniqueName: \"kubernetes.io/projected/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-kube-api-access-xl578\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.573842 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9e04f0b8-285c-4539-9996-9df4429c7bf8-horizon-secret-key\") pod \"9e04f0b8-285c-4539-9996-9df4429c7bf8\" (UID: \"9e04f0b8-285c-4539-9996-9df4429c7bf8\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.573897 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.573955 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/59e981b3-a120-4a1b-937c-8cc3892392de-config-data\") pod \"59e981b3-a120-4a1b-937c-8cc3892392de\" (UID: \"59e981b3-a120-4a1b-937c-8cc3892392de\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.574029 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-config-data\") pod \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.574057 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e04f0b8-285c-4539-9996-9df4429c7bf8-scripts\") pod \"9e04f0b8-285c-4539-9996-9df4429c7bf8\" (UID: \"9e04f0b8-285c-4539-9996-9df4429c7bf8\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.574077 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/59e981b3-a120-4a1b-937c-8cc3892392de-horizon-secret-key\") pod \"59e981b3-a120-4a1b-937c-8cc3892392de\" (UID: \"59e981b3-a120-4a1b-937c-8cc3892392de\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.574096 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-httpd-run\") pod \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.574713 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "07a92f93-a9f5-4c97-a903-2ab6d43a2ae4" (UID: "07a92f93-a9f5-4c97-a903-2ab6d43a2ae4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.574748 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e04f0b8-285c-4539-9996-9df4429c7bf8-scripts" (OuterVolumeSpecName: "scripts") pod "9e04f0b8-285c-4539-9996-9df4429c7bf8" (UID: "9e04f0b8-285c-4539-9996-9df4429c7bf8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.574863 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59e981b3-a120-4a1b-937c-8cc3892392de-config-data" (OuterVolumeSpecName: "config-data") pod "59e981b3-a120-4a1b-937c-8cc3892392de" (UID: "59e981b3-a120-4a1b-937c-8cc3892392de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.575198 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-config\") pod \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.575230 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmtwm\" (UniqueName: \"kubernetes.io/projected/9e04f0b8-285c-4539-9996-9df4429c7bf8-kube-api-access-kmtwm\") pod \"9e04f0b8-285c-4539-9996-9df4429c7bf8\" (UID: \"9e04f0b8-285c-4539-9996-9df4429c7bf8\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.575251 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-ovsdbserver-sb\") pod \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.575304 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-logs\") pod \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.575354 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-public-tls-certs\") pod \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.575376 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e04f0b8-285c-4539-9996-9df4429c7bf8-config-data\") pod \"9e04f0b8-285c-4539-9996-9df4429c7bf8\" (UID: \"9e04f0b8-285c-4539-9996-9df4429c7bf8\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.575427 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-combined-ca-bundle\") pod \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.575654 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-dns-svc\") pod \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.576264 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqslh\" (UniqueName: \"kubernetes.io/projected/59e981b3-a120-4a1b-937c-8cc3892392de-kube-api-access-wqslh\") pod \"59e981b3-a120-4a1b-937c-8cc3892392de\" (UID: \"59e981b3-a120-4a1b-937c-8cc3892392de\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.576321 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2z62x\" (UniqueName: \"kubernetes.io/projected/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-kube-api-access-2z62x\") pod \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.576345 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-ovsdbserver-nb\") pod \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.576375 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e04f0b8-285c-4539-9996-9df4429c7bf8-config-data" (OuterVolumeSpecName: "config-data") pod "9e04f0b8-285c-4539-9996-9df4429c7bf8" (UID: "9e04f0b8-285c-4539-9996-9df4429c7bf8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.576402 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/59e981b3-a120-4a1b-937c-8cc3892392de-scripts\") pod \"59e981b3-a120-4a1b-937c-8cc3892392de\" (UID: \"59e981b3-a120-4a1b-937c-8cc3892392de\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.576497 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59e981b3-a120-4a1b-937c-8cc3892392de-logs\") pod \"59e981b3-a120-4a1b-937c-8cc3892392de\" (UID: \"59e981b3-a120-4a1b-937c-8cc3892392de\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.577184 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-logs" (OuterVolumeSpecName: "logs") pod "07a92f93-a9f5-4c97-a903-2ab6d43a2ae4" (UID: "07a92f93-a9f5-4c97-a903-2ab6d43a2ae4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.577246 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59e981b3-a120-4a1b-937c-8cc3892392de-scripts" (OuterVolumeSpecName: "scripts") pod "59e981b3-a120-4a1b-937c-8cc3892392de" (UID: "59e981b3-a120-4a1b-937c-8cc3892392de"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.577453 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e04f0b8-285c-4539-9996-9df4429c7bf8-logs\") pod \"9e04f0b8-285c-4539-9996-9df4429c7bf8\" (UID: \"9e04f0b8-285c-4539-9996-9df4429c7bf8\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.577509 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7q6t\" (UniqueName: \"kubernetes.io/projected/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-kube-api-access-l7q6t\") pod \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.577549 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-scripts\") pod \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\" (UID: \"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.577566 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-dns-swift-storage-0\") pod \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\" (UID: \"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5\") " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.577833 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-logs\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.577935 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.577978 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.578095 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.578113 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.578164 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.578188 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.578304 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl578\" (UniqueName: \"kubernetes.io/projected/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-kube-api-access-xl578\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.578378 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/59e981b3-a120-4a1b-937c-8cc3892392de-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.578388 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e04f0b8-285c-4539-9996-9df4429c7bf8-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.578397 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.578405 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.578413 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e04f0b8-285c-4539-9996-9df4429c7bf8-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.578422 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/59e981b3-a120-4a1b-937c-8cc3892392de-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.578942 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59e981b3-a120-4a1b-937c-8cc3892392de-logs" (OuterVolumeSpecName: "logs") pod "59e981b3-a120-4a1b-937c-8cc3892392de" (UID: "59e981b3-a120-4a1b-937c-8cc3892392de"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.579161 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e04f0b8-285c-4539-9996-9df4429c7bf8-logs" (OuterVolumeSpecName: "logs") pod "9e04f0b8-285c-4539-9996-9df4429c7bf8" (UID: "9e04f0b8-285c-4539-9996-9df4429c7bf8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.579487 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e04f0b8-285c-4539-9996-9df4429c7bf8-kube-api-access-kmtwm" (OuterVolumeSpecName: "kube-api-access-kmtwm") pod "9e04f0b8-285c-4539-9996-9df4429c7bf8" (UID: "9e04f0b8-285c-4539-9996-9df4429c7bf8"). InnerVolumeSpecName "kube-api-access-kmtwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.580198 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e04f0b8-285c-4539-9996-9df4429c7bf8-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "9e04f0b8-285c-4539-9996-9df4429c7bf8" (UID: "9e04f0b8-285c-4539-9996-9df4429c7bf8"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.580493 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59e981b3-a120-4a1b-937c-8cc3892392de-kube-api-access-wqslh" (OuterVolumeSpecName: "kube-api-access-wqslh") pod "59e981b3-a120-4a1b-937c-8cc3892392de" (UID: "59e981b3-a120-4a1b-937c-8cc3892392de"). InnerVolumeSpecName "kube-api-access-wqslh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.582154 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59e981b3-a120-4a1b-937c-8cc3892392de-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "59e981b3-a120-4a1b-937c-8cc3892392de" (UID: "59e981b3-a120-4a1b-937c-8cc3892392de"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.582175 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-kube-api-access-2z62x" (OuterVolumeSpecName: "kube-api-access-2z62x") pod "07a92f93-a9f5-4c97-a903-2ab6d43a2ae4" (UID: "07a92f93-a9f5-4c97-a903-2ab6d43a2ae4"). InnerVolumeSpecName "kube-api-access-2z62x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.582573 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "07a92f93-a9f5-4c97-a903-2ab6d43a2ae4" (UID: "07a92f93-a9f5-4c97-a903-2ab6d43a2ae4"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.582671 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.584453 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-kube-api-access-l7q6t" (OuterVolumeSpecName: "kube-api-access-l7q6t") pod "1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5" (UID: "1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5"). InnerVolumeSpecName "kube-api-access-l7q6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.586840 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.586906 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.587136 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-logs\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.587391 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.589137 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.592042 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.598507 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl578\" (UniqueName: \"kubernetes.io/projected/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-kube-api-access-xl578\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.608882 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-scripts" (OuterVolumeSpecName: "scripts") pod "07a92f93-a9f5-4c97-a903-2ab6d43a2ae4" (UID: "07a92f93-a9f5-4c97-a903-2ab6d43a2ae4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.614134 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "07a92f93-a9f5-4c97-a903-2ab6d43a2ae4" (UID: "07a92f93-a9f5-4c97-a903-2ab6d43a2ae4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.625145 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.645450 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5" (UID: "1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.651156 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5" (UID: "1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.656493 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-config-data" (OuterVolumeSpecName: "config-data") pod "07a92f93-a9f5-4c97-a903-2ab6d43a2ae4" (UID: "07a92f93-a9f5-4c97-a903-2ab6d43a2ae4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.656953 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5" (UID: "1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.663042 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-config" (OuterVolumeSpecName: "config") pod "1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5" (UID: "1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.670312 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "07a92f93-a9f5-4c97-a903-2ab6d43a2ae4" (UID: "07a92f93-a9f5-4c97-a903-2ab6d43a2ae4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.679673 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.679707 4809 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/59e981b3-a120-4a1b-937c-8cc3892392de-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.679722 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.679737 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmtwm\" (UniqueName: \"kubernetes.io/projected/9e04f0b8-285c-4539-9996-9df4429c7bf8-kube-api-access-kmtwm\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.679749 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.679760 4809 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.679770 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.679781 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.679791 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqslh\" (UniqueName: \"kubernetes.io/projected/59e981b3-a120-4a1b-937c-8cc3892392de-kube-api-access-wqslh\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.679801 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2z62x\" (UniqueName: \"kubernetes.io/projected/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-kube-api-access-2z62x\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.679813 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.679825 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59e981b3-a120-4a1b-937c-8cc3892392de-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.679834 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e04f0b8-285c-4539-9996-9df4429c7bf8-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.679845 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7q6t\" (UniqueName: \"kubernetes.io/projected/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-kube-api-access-l7q6t\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.679854 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.679864 4809 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9e04f0b8-285c-4539-9996-9df4429c7bf8-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.679898 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.681145 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5" (UID: "1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.684554 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.697205 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.781408 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.781443 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:04 crc kubenswrapper[4809]: I1124 07:12:04.955629 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1" path="/var/lib/kubelet/pods/b21d4e20-6b98-49a9-9e6f-02bb8d88c7b1/volumes" Nov 24 07:12:05 crc kubenswrapper[4809]: E1124 07:12:05.216486 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e04f0b8_285c_4539_9996_9df4429c7bf8.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e04f0b8_285c_4539_9996_9df4429c7bf8.slice/crio-dd3f12cb214167bc81f8dede51e2a991e120821ebc16b856e61021e3b0936c51\": RecentStats: unable to find data in memory cache]" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.304618 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-56545965c7-wbhlj" event={"ID":"59e981b3-a120-4a1b-937c-8cc3892392de","Type":"ContainerDied","Data":"92f4ed3d551452ef2b1e4c77f0f367d7059974d41f63b70cdbbcd81b18d95a13"} Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.304709 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-56545965c7-wbhlj" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.310858 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.311048 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" event={"ID":"1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5","Type":"ContainerDied","Data":"0da44b051b38ed2c65e0b9a2118051637203c5bff074f9eb1e37b02528fec3fa"} Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.315157 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"07a92f93-a9f5-4c97-a903-2ab6d43a2ae4","Type":"ContainerDied","Data":"459a35f6f64d83ee8ee13446cf9e3f7a6f1adf176576b930c284de24887afb8d"} Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.315251 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.321584 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-546d49d8b5-77xgm" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.322220 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-546d49d8b5-77xgm" event={"ID":"9e04f0b8-285c-4539-9996-9df4429c7bf8","Type":"ContainerDied","Data":"dd3f12cb214167bc81f8dede51e2a991e120821ebc16b856e61021e3b0936c51"} Nov 24 07:12:05 crc kubenswrapper[4809]: E1124 07:12:05.324107 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-lkj6z" podUID="972cb0ce-ea08-408c-a10d-ba1d14d42568" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.381700 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-56545965c7-wbhlj"] Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.403501 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-56545965c7-wbhlj"] Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.431241 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.447089 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.447145 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-2b7fh"] Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.455455 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-2b7fh"] Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.458542 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:12:05 crc kubenswrapper[4809]: E1124 07:12:05.458896 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07a92f93-a9f5-4c97-a903-2ab6d43a2ae4" containerName="glance-httpd" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.458911 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="07a92f93-a9f5-4c97-a903-2ab6d43a2ae4" containerName="glance-httpd" Nov 24 07:12:05 crc kubenswrapper[4809]: E1124 07:12:05.458936 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07a92f93-a9f5-4c97-a903-2ab6d43a2ae4" containerName="glance-log" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.458942 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="07a92f93-a9f5-4c97-a903-2ab6d43a2ae4" containerName="glance-log" Nov 24 07:12:05 crc kubenswrapper[4809]: E1124 07:12:05.458954 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5" containerName="dnsmasq-dns" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.458959 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5" containerName="dnsmasq-dns" Nov 24 07:12:05 crc kubenswrapper[4809]: E1124 07:12:05.458991 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5" containerName="init" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.458997 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5" containerName="init" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.459185 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5" containerName="dnsmasq-dns" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.459208 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="07a92f93-a9f5-4c97-a903-2ab6d43a2ae4" containerName="glance-log" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.459224 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="07a92f93-a9f5-4c97-a903-2ab6d43a2ae4" containerName="glance-httpd" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.460127 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.463311 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.463470 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.469613 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.491842 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-546d49d8b5-77xgm"] Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.505972 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-546d49d8b5-77xgm"] Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.596474 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-scripts\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.596528 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29842a6a-e532-4757-829e-418ecc738eca-logs\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.596557 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.596591 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/29842a6a-e532-4757-829e-418ecc738eca-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.596619 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.596642 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.596663 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-config-data\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.596725 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgzbl\" (UniqueName: \"kubernetes.io/projected/29842a6a-e532-4757-829e-418ecc738eca-kube-api-access-pgzbl\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.698054 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-scripts\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.698397 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29842a6a-e532-4757-829e-418ecc738eca-logs\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.698435 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.698467 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/29842a6a-e532-4757-829e-418ecc738eca-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.698489 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.698506 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.698528 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-config-data\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.698594 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgzbl\" (UniqueName: \"kubernetes.io/projected/29842a6a-e532-4757-829e-418ecc738eca-kube-api-access-pgzbl\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.699241 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.699268 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29842a6a-e532-4757-829e-418ecc738eca-logs\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.700903 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/29842a6a-e532-4757-829e-418ecc738eca-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.704629 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.717832 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-scripts\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.718031 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgzbl\" (UniqueName: \"kubernetes.io/projected/29842a6a-e532-4757-829e-418ecc738eca-kube-api-access-pgzbl\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.718772 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.720047 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-config-data\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.735147 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " pod="openstack/glance-default-external-api-0" Nov 24 07:12:05 crc kubenswrapper[4809]: I1124 07:12:05.777056 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 07:12:06 crc kubenswrapper[4809]: E1124 07:12:06.127623 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 24 07:12:06 crc kubenswrapper[4809]: E1124 07:12:06.127777 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9q8n6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-dtg82_openstack(35dbbf41-295e-4ec9-bd02-12911c093ce8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 07:12:06 crc kubenswrapper[4809]: E1124 07:12:06.129473 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-dtg82" podUID="35dbbf41-295e-4ec9-bd02-12911c093ce8" Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.162432 4809 scope.go:117] "RemoveContainer" containerID="eeca6acca22917a6233a98d4cb48281b10499512c51d94c995a6831ab1332f5f" Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.308502 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4l5jw" Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.335734 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4l5jw" Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.335812 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4l5jw" event={"ID":"8cc0b19f-d90f-4460-909e-6eb644cb2955","Type":"ContainerDied","Data":"7144dc03fe2711c0989bb80d008b55405d3366cfffa8d2ee9a89ff62a35d8da8"} Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.335851 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7144dc03fe2711c0989bb80d008b55405d3366cfffa8d2ee9a89ff62a35d8da8" Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.368466 4809 scope.go:117] "RemoveContainer" containerID="95db14eaac6f10051b603ed6695d3a7360907b8e405a445d744910d120259bbe" Nov 24 07:12:06 crc kubenswrapper[4809]: E1124 07:12:06.368579 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-dtg82" podUID="35dbbf41-295e-4ec9-bd02-12911c093ce8" Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.408469 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cc0b19f-d90f-4460-909e-6eb644cb2955-combined-ca-bundle\") pod \"8cc0b19f-d90f-4460-909e-6eb644cb2955\" (UID: \"8cc0b19f-d90f-4460-909e-6eb644cb2955\") " Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.408834 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8cc0b19f-d90f-4460-909e-6eb644cb2955-config\") pod \"8cc0b19f-d90f-4460-909e-6eb644cb2955\" (UID: \"8cc0b19f-d90f-4460-909e-6eb644cb2955\") " Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.408955 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8frz\" (UniqueName: \"kubernetes.io/projected/8cc0b19f-d90f-4460-909e-6eb644cb2955-kube-api-access-t8frz\") pod \"8cc0b19f-d90f-4460-909e-6eb644cb2955\" (UID: \"8cc0b19f-d90f-4460-909e-6eb644cb2955\") " Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.415110 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cc0b19f-d90f-4460-909e-6eb644cb2955-kube-api-access-t8frz" (OuterVolumeSpecName: "kube-api-access-t8frz") pod "8cc0b19f-d90f-4460-909e-6eb644cb2955" (UID: "8cc0b19f-d90f-4460-909e-6eb644cb2955"). InnerVolumeSpecName "kube-api-access-t8frz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.415653 4809 scope.go:117] "RemoveContainer" containerID="056e7b7f88d76dbaaa1c9ceaa438598701da283180eba591690d9e0dba040281" Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.440153 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cc0b19f-d90f-4460-909e-6eb644cb2955-config" (OuterVolumeSpecName: "config") pod "8cc0b19f-d90f-4460-909e-6eb644cb2955" (UID: "8cc0b19f-d90f-4460-909e-6eb644cb2955"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.443346 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cc0b19f-d90f-4460-909e-6eb644cb2955-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8cc0b19f-d90f-4460-909e-6eb644cb2955" (UID: "8cc0b19f-d90f-4460-909e-6eb644cb2955"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.463831 4809 scope.go:117] "RemoveContainer" containerID="07b8047b10191667e2554f1262e98edd9ac90565c43aa23e76a0a199917caa80" Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.496776 4809 scope.go:117] "RemoveContainer" containerID="b233c6ea9c0134a2310d53352dd483c95217c6794362c9c5dc9b25069e495e3e" Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.511481 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cc0b19f-d90f-4460-909e-6eb644cb2955-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.511513 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8cc0b19f-d90f-4460-909e-6eb644cb2955-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.511523 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8frz\" (UniqueName: \"kubernetes.io/projected/8cc0b19f-d90f-4460-909e-6eb644cb2955-kube-api-access-t8frz\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.572728 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8585547888-qdlwv"] Nov 24 07:12:06 crc kubenswrapper[4809]: W1124 07:12:06.579125 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5487494_236a_4335_977f_11167eaa3cdc.slice/crio-a9e3619fd3a30f5208f3edfa581fd8067925cdf13c96e76f1c83271d3b0c1291 WatchSource:0}: Error finding container a9e3619fd3a30f5208f3edfa581fd8067925cdf13c96e76f1c83271d3b0c1291: Status 404 returned error can't find the container with id a9e3619fd3a30f5208f3edfa581fd8067925cdf13c96e76f1c83271d3b0c1291 Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.728531 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-qsvl5"] Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.733701 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79cbd445d4-5hkc8"] Nov 24 07:12:06 crc kubenswrapper[4809]: W1124 07:12:06.734131 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5d93390_ea3b_43d3_83ba_4d85fb8544d4.slice/crio-ec7caacc262d210a11c5099e18223026361c4d4be972849ac7376348764700bf WatchSource:0}: Error finding container ec7caacc262d210a11c5099e18223026361c4d4be972849ac7376348764700bf: Status 404 returned error can't find the container with id ec7caacc262d210a11c5099e18223026361c4d4be972849ac7376348764700bf Nov 24 07:12:06 crc kubenswrapper[4809]: W1124 07:12:06.734955 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b3e7ca6_2814_4a14_b1f4_765f8086927e.slice/crio-a8a153a7299d3120e11605db7a08ba8650c559c4993bce9f50a9b6d6d11a53cd WatchSource:0}: Error finding container a8a153a7299d3120e11605db7a08ba8650c559c4993bce9f50a9b6d6d11a53cd: Status 404 returned error can't find the container with id a8a153a7299d3120e11605db7a08ba8650c559c4993bce9f50a9b6d6d11a53cd Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.738902 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.901919 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07a92f93-a9f5-4c97-a903-2ab6d43a2ae4" path="/var/lib/kubelet/pods/07a92f93-a9f5-4c97-a903-2ab6d43a2ae4/volumes" Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.905192 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5" path="/var/lib/kubelet/pods/1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5/volumes" Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.905823 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59e981b3-a120-4a1b-937c-8cc3892392de" path="/var/lib/kubelet/pods/59e981b3-a120-4a1b-937c-8cc3892392de/volumes" Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.906237 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e04f0b8-285c-4539-9996-9df4429c7bf8" path="/var/lib/kubelet/pods/9e04f0b8-285c-4539-9996-9df4429c7bf8/volumes" Nov 24 07:12:06 crc kubenswrapper[4809]: I1124 07:12:06.909326 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:12:06 crc kubenswrapper[4809]: W1124 07:12:06.915815 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0bf846c6_e717_4fab_8fe0_0d43d1747ce4.slice/crio-f82b340d02e56520c08a4ed546872cc9bdce4cf3ada4b594d0987107ca2bc1bf WatchSource:0}: Error finding container f82b340d02e56520c08a4ed546872cc9bdce4cf3ada4b594d0987107ca2bc1bf: Status 404 returned error can't find the container with id f82b340d02e56520c08a4ed546872cc9bdce4cf3ada4b594d0987107ca2bc1bf Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.350853 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79cbd445d4-5hkc8" event={"ID":"5b3e7ca6-2814-4a14-b1f4-765f8086927e","Type":"ContainerStarted","Data":"54e0622a23539be0abf29e32ee3cbc72076ffc23105f373a61b7e155e5acb13f"} Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.351138 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79cbd445d4-5hkc8" event={"ID":"5b3e7ca6-2814-4a14-b1f4-765f8086927e","Type":"ContainerStarted","Data":"1a68297f5bbb4fbec6ca1cf7ed7ec3aaee4d1d4d2c979776b78c71e5790046cc"} Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.351148 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79cbd445d4-5hkc8" event={"ID":"5b3e7ca6-2814-4a14-b1f4-765f8086927e","Type":"ContainerStarted","Data":"a8a153a7299d3120e11605db7a08ba8650c559c4993bce9f50a9b6d6d11a53cd"} Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.369269 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-79cbd445d4-5hkc8" podStartSLOduration=23.369251849 podStartE2EDuration="23.369251849s" podCreationTimestamp="2025-11-24 07:11:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:12:07.366433324 +0000 UTC m=+1027.267025029" watchObservedRunningTime="2025-11-24 07:12:07.369251849 +0000 UTC m=+1027.269843554" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.379093 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8585547888-qdlwv" event={"ID":"b5487494-236a-4335-977f-11167eaa3cdc","Type":"ContainerStarted","Data":"db423d7a206d0c11ae39868d2da70a65fc53c4a40242017f01566ff7c901face"} Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.379150 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8585547888-qdlwv" event={"ID":"b5487494-236a-4335-977f-11167eaa3cdc","Type":"ContainerStarted","Data":"ae4fa4e44ef01baf252d001f019090c77e7d687e7d93f74a7318583e1e1967fa"} Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.379162 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8585547888-qdlwv" event={"ID":"b5487494-236a-4335-977f-11167eaa3cdc","Type":"ContainerStarted","Data":"a9e3619fd3a30f5208f3edfa581fd8067925cdf13c96e76f1c83271d3b0c1291"} Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.385188 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qsvl5" event={"ID":"f5d93390-ea3b-43d3-83ba-4d85fb8544d4","Type":"ContainerStarted","Data":"01ef821104ff2fbdbeaf37ceb28bf62950efaea3723a70ac8cfabc611a5820a4"} Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.385232 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qsvl5" event={"ID":"f5d93390-ea3b-43d3-83ba-4d85fb8544d4","Type":"ContainerStarted","Data":"ec7caacc262d210a11c5099e18223026361c4d4be972849ac7376348764700bf"} Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.390237 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"655e6590-673e-4fde-a097-1173dfcef9e2","Type":"ContainerStarted","Data":"42f9be194d69131769e788b4521d542fb8b90d88ac5fe9d2fc53be376177c837"} Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.392838 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84666554f7-jc9pp" event={"ID":"46ebff1a-add8-420f-99c2-c593d486d2f8","Type":"ContainerStarted","Data":"e44083aa1ad4fa4b45a13e3c4271aa1d170097c126ada26dcc06b313160fc1c8"} Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.392894 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84666554f7-jc9pp" event={"ID":"46ebff1a-add8-420f-99c2-c593d486d2f8","Type":"ContainerStarted","Data":"31646cbdae7ff94e4b915bd134d0575d662c1ea9527419ba1ba39e2845859955"} Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.393019 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-84666554f7-jc9pp" podUID="46ebff1a-add8-420f-99c2-c593d486d2f8" containerName="horizon-log" containerID="cri-o://31646cbdae7ff94e4b915bd134d0575d662c1ea9527419ba1ba39e2845859955" gracePeriod=30 Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.393265 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-84666554f7-jc9pp" podUID="46ebff1a-add8-420f-99c2-c593d486d2f8" containerName="horizon" containerID="cri-o://e44083aa1ad4fa4b45a13e3c4271aa1d170097c126ada26dcc06b313160fc1c8" gracePeriod=30 Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.406923 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-8585547888-qdlwv" podStartSLOduration=23.406899518 podStartE2EDuration="23.406899518s" podCreationTimestamp="2025-11-24 07:11:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:12:07.400717614 +0000 UTC m=+1027.301309319" watchObservedRunningTime="2025-11-24 07:12:07.406899518 +0000 UTC m=+1027.307491223" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.417641 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0bf846c6-e717-4fab-8fe0-0d43d1747ce4","Type":"ContainerStarted","Data":"f82b340d02e56520c08a4ed546872cc9bdce4cf3ada4b594d0987107ca2bc1bf"} Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.435305 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-qsvl5" podStartSLOduration=15.43526106 podStartE2EDuration="15.43526106s" podCreationTimestamp="2025-11-24 07:11:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:12:07.419599594 +0000 UTC m=+1027.320191299" watchObservedRunningTime="2025-11-24 07:12:07.43526106 +0000 UTC m=+1027.335852765" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.450714 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-84666554f7-jc9pp" podStartSLOduration=4.083565178 podStartE2EDuration="28.450697749s" podCreationTimestamp="2025-11-24 07:11:39 +0000 UTC" firstStartedPulling="2025-11-24 07:11:40.062461563 +0000 UTC m=+999.963053268" lastFinishedPulling="2025-11-24 07:12:04.429594134 +0000 UTC m=+1024.330185839" observedRunningTime="2025-11-24 07:12:07.43867751 +0000 UTC m=+1027.339269225" watchObservedRunningTime="2025-11-24 07:12:07.450697749 +0000 UTC m=+1027.351289454" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.577987 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-2pm9c"] Nov 24 07:12:07 crc kubenswrapper[4809]: E1124 07:12:07.578366 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cc0b19f-d90f-4460-909e-6eb644cb2955" containerName="neutron-db-sync" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.578384 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cc0b19f-d90f-4460-909e-6eb644cb2955" containerName="neutron-db-sync" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.578567 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cc0b19f-d90f-4460-909e-6eb644cb2955" containerName="neutron-db-sync" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.579408 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.601929 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-2pm9c"] Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.648177 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-68b795db4d-rr6j7"] Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.649951 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-68b795db4d-rr6j7" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.659295 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.659924 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.660202 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-qgc8p" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.661657 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.668022 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-68b795db4d-rr6j7"] Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.736252 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-ovndb-tls-certs\") pod \"neutron-68b795db4d-rr6j7\" (UID: \"923046a7-b198-408c-a853-0c54d8adc84c\") " pod="openstack/neutron-68b795db4d-rr6j7" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.736313 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-config\") pod \"dnsmasq-dns-55f844cf75-2pm9c\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.736350 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-dns-svc\") pod \"dnsmasq-dns-55f844cf75-2pm9c\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.736408 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhhjr\" (UniqueName: \"kubernetes.io/projected/923046a7-b198-408c-a853-0c54d8adc84c-kube-api-access-hhhjr\") pod \"neutron-68b795db4d-rr6j7\" (UID: \"923046a7-b198-408c-a853-0c54d8adc84c\") " pod="openstack/neutron-68b795db4d-rr6j7" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.736428 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zls8\" (UniqueName: \"kubernetes.io/projected/07b5d297-b3f9-4b51-a007-045cec29345f-kube-api-access-5zls8\") pod \"dnsmasq-dns-55f844cf75-2pm9c\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.736472 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-config\") pod \"neutron-68b795db4d-rr6j7\" (UID: \"923046a7-b198-408c-a853-0c54d8adc84c\") " pod="openstack/neutron-68b795db4d-rr6j7" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.736504 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-2pm9c\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.736518 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-httpd-config\") pod \"neutron-68b795db4d-rr6j7\" (UID: \"923046a7-b198-408c-a853-0c54d8adc84c\") " pod="openstack/neutron-68b795db4d-rr6j7" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.736532 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-2pm9c\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.736556 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-combined-ca-bundle\") pod \"neutron-68b795db4d-rr6j7\" (UID: \"923046a7-b198-408c-a853-0c54d8adc84c\") " pod="openstack/neutron-68b795db4d-rr6j7" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.736582 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-2pm9c\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.771459 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.837553 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-2pm9c\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.837644 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-ovndb-tls-certs\") pod \"neutron-68b795db4d-rr6j7\" (UID: \"923046a7-b198-408c-a853-0c54d8adc84c\") " pod="openstack/neutron-68b795db4d-rr6j7" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.837668 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-config\") pod \"dnsmasq-dns-55f844cf75-2pm9c\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.837694 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-dns-svc\") pod \"dnsmasq-dns-55f844cf75-2pm9c\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.837719 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhhjr\" (UniqueName: \"kubernetes.io/projected/923046a7-b198-408c-a853-0c54d8adc84c-kube-api-access-hhhjr\") pod \"neutron-68b795db4d-rr6j7\" (UID: \"923046a7-b198-408c-a853-0c54d8adc84c\") " pod="openstack/neutron-68b795db4d-rr6j7" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.837741 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zls8\" (UniqueName: \"kubernetes.io/projected/07b5d297-b3f9-4b51-a007-045cec29345f-kube-api-access-5zls8\") pod \"dnsmasq-dns-55f844cf75-2pm9c\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.837767 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-config\") pod \"neutron-68b795db4d-rr6j7\" (UID: \"923046a7-b198-408c-a853-0c54d8adc84c\") " pod="openstack/neutron-68b795db4d-rr6j7" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.837801 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-2pm9c\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.837819 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-httpd-config\") pod \"neutron-68b795db4d-rr6j7\" (UID: \"923046a7-b198-408c-a853-0c54d8adc84c\") " pod="openstack/neutron-68b795db4d-rr6j7" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.837839 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-2pm9c\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.837867 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-combined-ca-bundle\") pod \"neutron-68b795db4d-rr6j7\" (UID: \"923046a7-b198-408c-a853-0c54d8adc84c\") " pod="openstack/neutron-68b795db4d-rr6j7" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.839621 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-2pm9c\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.839709 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-config\") pod \"dnsmasq-dns-55f844cf75-2pm9c\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.840184 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-dns-svc\") pod \"dnsmasq-dns-55f844cf75-2pm9c\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.840740 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-2pm9c\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.841508 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-2pm9c\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.845396 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-ovndb-tls-certs\") pod \"neutron-68b795db4d-rr6j7\" (UID: \"923046a7-b198-408c-a853-0c54d8adc84c\") " pod="openstack/neutron-68b795db4d-rr6j7" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.846744 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-combined-ca-bundle\") pod \"neutron-68b795db4d-rr6j7\" (UID: \"923046a7-b198-408c-a853-0c54d8adc84c\") " pod="openstack/neutron-68b795db4d-rr6j7" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.856371 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-httpd-config\") pod \"neutron-68b795db4d-rr6j7\" (UID: \"923046a7-b198-408c-a853-0c54d8adc84c\") " pod="openstack/neutron-68b795db4d-rr6j7" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.857753 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-config\") pod \"neutron-68b795db4d-rr6j7\" (UID: \"923046a7-b198-408c-a853-0c54d8adc84c\") " pod="openstack/neutron-68b795db4d-rr6j7" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.857927 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zls8\" (UniqueName: \"kubernetes.io/projected/07b5d297-b3f9-4b51-a007-045cec29345f-kube-api-access-5zls8\") pod \"dnsmasq-dns-55f844cf75-2pm9c\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.858658 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhhjr\" (UniqueName: \"kubernetes.io/projected/923046a7-b198-408c-a853-0c54d8adc84c-kube-api-access-hhhjr\") pod \"neutron-68b795db4d-rr6j7\" (UID: \"923046a7-b198-408c-a853-0c54d8adc84c\") " pod="openstack/neutron-68b795db4d-rr6j7" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.921909 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:07 crc kubenswrapper[4809]: I1124 07:12:07.977930 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-68b795db4d-rr6j7" Nov 24 07:12:08 crc kubenswrapper[4809]: W1124 07:12:08.140078 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod29842a6a_e532_4757_829e_418ecc738eca.slice/crio-dbba0e81948331aafcc04809174f2e9a3efe2000e5fbbc136048d5e9435da9a5 WatchSource:0}: Error finding container dbba0e81948331aafcc04809174f2e9a3efe2000e5fbbc136048d5e9435da9a5: Status 404 returned error can't find the container with id dbba0e81948331aafcc04809174f2e9a3efe2000e5fbbc136048d5e9435da9a5 Nov 24 07:12:08 crc kubenswrapper[4809]: I1124 07:12:08.454137 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"29842a6a-e532-4757-829e-418ecc738eca","Type":"ContainerStarted","Data":"dbba0e81948331aafcc04809174f2e9a3efe2000e5fbbc136048d5e9435da9a5"} Nov 24 07:12:08 crc kubenswrapper[4809]: I1124 07:12:08.481908 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0bf846c6-e717-4fab-8fe0-0d43d1747ce4","Type":"ContainerStarted","Data":"b15a6958eff3edbf8366b4ad13629375e82ded4b11e08ed984f470b7add61f0e"} Nov 24 07:12:08 crc kubenswrapper[4809]: I1124 07:12:08.771029 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-68b795db4d-rr6j7"] Nov 24 07:12:08 crc kubenswrapper[4809]: W1124 07:12:08.781073 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod923046a7_b198_408c_a853_0c54d8adc84c.slice/crio-0d0abc01fe8dad3163a0732353cd8ebc107da004248fbfeb459246573f397e6c WatchSource:0}: Error finding container 0d0abc01fe8dad3163a0732353cd8ebc107da004248fbfeb459246573f397e6c: Status 404 returned error can't find the container with id 0d0abc01fe8dad3163a0732353cd8ebc107da004248fbfeb459246573f397e6c Nov 24 07:12:08 crc kubenswrapper[4809]: I1124 07:12:08.818039 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-2b7fh" podUID="1c30be2a-1a3c-41c2-95fc-1f1ad8e23ed5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.131:5353: i/o timeout" Nov 24 07:12:08 crc kubenswrapper[4809]: I1124 07:12:08.842206 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-2pm9c"] Nov 24 07:12:09 crc kubenswrapper[4809]: I1124 07:12:09.484531 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-84666554f7-jc9pp" Nov 24 07:12:09 crc kubenswrapper[4809]: I1124 07:12:09.491442 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68b795db4d-rr6j7" event={"ID":"923046a7-b198-408c-a853-0c54d8adc84c","Type":"ContainerStarted","Data":"c7abafbc9de95f0d2c23e57deab62dd208867121a2e1920e34a4c6dfde6251b2"} Nov 24 07:12:09 crc kubenswrapper[4809]: I1124 07:12:09.491472 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68b795db4d-rr6j7" event={"ID":"923046a7-b198-408c-a853-0c54d8adc84c","Type":"ContainerStarted","Data":"0d0abc01fe8dad3163a0732353cd8ebc107da004248fbfeb459246573f397e6c"} Nov 24 07:12:09 crc kubenswrapper[4809]: I1124 07:12:09.494585 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"655e6590-673e-4fde-a097-1173dfcef9e2","Type":"ContainerStarted","Data":"ffc5ddd380aa10bc65340c8e0facbe2c23a4aba66abadd3808bb97c90888fe85"} Nov 24 07:12:09 crc kubenswrapper[4809]: I1124 07:12:09.496044 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"29842a6a-e532-4757-829e-418ecc738eca","Type":"ContainerStarted","Data":"d606986fa98397e571fbd9ac327ccc96cd4c2550e7f041c4a3383d845df59ee0"} Nov 24 07:12:09 crc kubenswrapper[4809]: I1124 07:12:09.497848 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0bf846c6-e717-4fab-8fe0-0d43d1747ce4","Type":"ContainerStarted","Data":"ecc506aaa274fb6334ab278f177ffc8f8b8f2706a9ac605330d06ff15b1cd040"} Nov 24 07:12:09 crc kubenswrapper[4809]: I1124 07:12:09.499448 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" event={"ID":"07b5d297-b3f9-4b51-a007-045cec29345f","Type":"ContainerStarted","Data":"a208d8e30cccfa4caedd94d326730c550e78d8541611b8a3823638e6f99ef7be"} Nov 24 07:12:09 crc kubenswrapper[4809]: I1124 07:12:09.499482 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" event={"ID":"07b5d297-b3f9-4b51-a007-045cec29345f","Type":"ContainerStarted","Data":"2e04f5e4b004a048393f233666ce3f0bc76687bb542ed18865168521362cad8a"} Nov 24 07:12:09 crc kubenswrapper[4809]: I1124 07:12:09.535770 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.535754838 podStartE2EDuration="5.535754838s" podCreationTimestamp="2025-11-24 07:12:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:12:09.53434337 +0000 UTC m=+1029.434935075" watchObservedRunningTime="2025-11-24 07:12:09.535754838 +0000 UTC m=+1029.436346543" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.024044 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6c5d67c78c-vvvv9"] Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.028091 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.030811 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.031852 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.041579 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6c5d67c78c-vvvv9"] Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.090628 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c53f941-5f7d-489c-85eb-d07ff8d4ef6b-ovndb-tls-certs\") pod \"neutron-6c5d67c78c-vvvv9\" (UID: \"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b\") " pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.090696 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9c53f941-5f7d-489c-85eb-d07ff8d4ef6b-httpd-config\") pod \"neutron-6c5d67c78c-vvvv9\" (UID: \"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b\") " pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.090761 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c53f941-5f7d-489c-85eb-d07ff8d4ef6b-public-tls-certs\") pod \"neutron-6c5d67c78c-vvvv9\" (UID: \"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b\") " pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.090789 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9c53f941-5f7d-489c-85eb-d07ff8d4ef6b-config\") pod \"neutron-6c5d67c78c-vvvv9\" (UID: \"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b\") " pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.090860 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c53f941-5f7d-489c-85eb-d07ff8d4ef6b-combined-ca-bundle\") pod \"neutron-6c5d67c78c-vvvv9\" (UID: \"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b\") " pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.090893 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c53f941-5f7d-489c-85eb-d07ff8d4ef6b-internal-tls-certs\") pod \"neutron-6c5d67c78c-vvvv9\" (UID: \"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b\") " pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.090916 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sj64m\" (UniqueName: \"kubernetes.io/projected/9c53f941-5f7d-489c-85eb-d07ff8d4ef6b-kube-api-access-sj64m\") pod \"neutron-6c5d67c78c-vvvv9\" (UID: \"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b\") " pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.192374 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c53f941-5f7d-489c-85eb-d07ff8d4ef6b-internal-tls-certs\") pod \"neutron-6c5d67c78c-vvvv9\" (UID: \"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b\") " pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.192414 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sj64m\" (UniqueName: \"kubernetes.io/projected/9c53f941-5f7d-489c-85eb-d07ff8d4ef6b-kube-api-access-sj64m\") pod \"neutron-6c5d67c78c-vvvv9\" (UID: \"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b\") " pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.192495 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c53f941-5f7d-489c-85eb-d07ff8d4ef6b-ovndb-tls-certs\") pod \"neutron-6c5d67c78c-vvvv9\" (UID: \"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b\") " pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.192537 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9c53f941-5f7d-489c-85eb-d07ff8d4ef6b-httpd-config\") pod \"neutron-6c5d67c78c-vvvv9\" (UID: \"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b\") " pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.192604 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c53f941-5f7d-489c-85eb-d07ff8d4ef6b-public-tls-certs\") pod \"neutron-6c5d67c78c-vvvv9\" (UID: \"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b\") " pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.193552 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9c53f941-5f7d-489c-85eb-d07ff8d4ef6b-config\") pod \"neutron-6c5d67c78c-vvvv9\" (UID: \"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b\") " pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.193624 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c53f941-5f7d-489c-85eb-d07ff8d4ef6b-combined-ca-bundle\") pod \"neutron-6c5d67c78c-vvvv9\" (UID: \"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b\") " pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.202563 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c53f941-5f7d-489c-85eb-d07ff8d4ef6b-internal-tls-certs\") pod \"neutron-6c5d67c78c-vvvv9\" (UID: \"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b\") " pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.202591 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c53f941-5f7d-489c-85eb-d07ff8d4ef6b-ovndb-tls-certs\") pod \"neutron-6c5d67c78c-vvvv9\" (UID: \"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b\") " pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.202653 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c53f941-5f7d-489c-85eb-d07ff8d4ef6b-combined-ca-bundle\") pod \"neutron-6c5d67c78c-vvvv9\" (UID: \"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b\") " pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.203293 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9c53f941-5f7d-489c-85eb-d07ff8d4ef6b-config\") pod \"neutron-6c5d67c78c-vvvv9\" (UID: \"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b\") " pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.205188 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c53f941-5f7d-489c-85eb-d07ff8d4ef6b-public-tls-certs\") pod \"neutron-6c5d67c78c-vvvv9\" (UID: \"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b\") " pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.212712 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9c53f941-5f7d-489c-85eb-d07ff8d4ef6b-httpd-config\") pod \"neutron-6c5d67c78c-vvvv9\" (UID: \"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b\") " pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.214510 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sj64m\" (UniqueName: \"kubernetes.io/projected/9c53f941-5f7d-489c-85eb-d07ff8d4ef6b-kube-api-access-sj64m\") pod \"neutron-6c5d67c78c-vvvv9\" (UID: \"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b\") " pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.382673 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.510053 4809 generic.go:334] "Generic (PLEG): container finished" podID="07b5d297-b3f9-4b51-a007-045cec29345f" containerID="a208d8e30cccfa4caedd94d326730c550e78d8541611b8a3823638e6f99ef7be" exitCode=0 Nov 24 07:12:10 crc kubenswrapper[4809]: I1124 07:12:10.510885 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" event={"ID":"07b5d297-b3f9-4b51-a007-045cec29345f","Type":"ContainerDied","Data":"a208d8e30cccfa4caedd94d326730c550e78d8541611b8a3823638e6f99ef7be"} Nov 24 07:12:11 crc kubenswrapper[4809]: I1124 07:12:11.548738 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68b795db4d-rr6j7" event={"ID":"923046a7-b198-408c-a853-0c54d8adc84c","Type":"ContainerStarted","Data":"04f0aa5a4dd17cacb5c047a936950427bc739864b143d2d38607b108eb4a6f83"} Nov 24 07:12:11 crc kubenswrapper[4809]: I1124 07:12:11.549094 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-68b795db4d-rr6j7" Nov 24 07:12:11 crc kubenswrapper[4809]: I1124 07:12:11.569379 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-68b795db4d-rr6j7" podStartSLOduration=4.569363673 podStartE2EDuration="4.569363673s" podCreationTimestamp="2025-11-24 07:12:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:12:11.567306578 +0000 UTC m=+1031.467898283" watchObservedRunningTime="2025-11-24 07:12:11.569363673 +0000 UTC m=+1031.469955378" Nov 24 07:12:12 crc kubenswrapper[4809]: I1124 07:12:12.563803 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" event={"ID":"07b5d297-b3f9-4b51-a007-045cec29345f","Type":"ContainerStarted","Data":"d26d77e49721fe7649c0e65807e4597d375dff5e87e3630c335e62e5618c9a30"} Nov 24 07:12:12 crc kubenswrapper[4809]: I1124 07:12:12.564289 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:12 crc kubenswrapper[4809]: I1124 07:12:12.566930 4809 generic.go:334] "Generic (PLEG): container finished" podID="f5d93390-ea3b-43d3-83ba-4d85fb8544d4" containerID="01ef821104ff2fbdbeaf37ceb28bf62950efaea3723a70ac8cfabc611a5820a4" exitCode=0 Nov 24 07:12:12 crc kubenswrapper[4809]: I1124 07:12:12.567002 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qsvl5" event={"ID":"f5d93390-ea3b-43d3-83ba-4d85fb8544d4","Type":"ContainerDied","Data":"01ef821104ff2fbdbeaf37ceb28bf62950efaea3723a70ac8cfabc611a5820a4"} Nov 24 07:12:12 crc kubenswrapper[4809]: I1124 07:12:12.569169 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"29842a6a-e532-4757-829e-418ecc738eca","Type":"ContainerStarted","Data":"abb9fd0265dd261d032eb959dd6b9eb41c5e9d55f981ba63a0a7413c61c584a9"} Nov 24 07:12:12 crc kubenswrapper[4809]: I1124 07:12:12.587553 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6c5d67c78c-vvvv9"] Nov 24 07:12:12 crc kubenswrapper[4809]: I1124 07:12:12.588655 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" podStartSLOduration=5.588637516 podStartE2EDuration="5.588637516s" podCreationTimestamp="2025-11-24 07:12:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:12:12.586736995 +0000 UTC m=+1032.487328700" watchObservedRunningTime="2025-11-24 07:12:12.588637516 +0000 UTC m=+1032.489229221" Nov 24 07:12:12 crc kubenswrapper[4809]: I1124 07:12:12.606705 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.606683814 podStartE2EDuration="7.606683814s" podCreationTimestamp="2025-11-24 07:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:12:12.603277824 +0000 UTC m=+1032.503869529" watchObservedRunningTime="2025-11-24 07:12:12.606683814 +0000 UTC m=+1032.507275519" Nov 24 07:12:13 crc kubenswrapper[4809]: I1124 07:12:13.583341 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ncjd6" event={"ID":"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293","Type":"ContainerStarted","Data":"0d7c52cf5c5cc03d5a34b01cf1a3e2ef1ec36b4324033c502207eeed4da969fa"} Nov 24 07:12:13 crc kubenswrapper[4809]: I1124 07:12:13.588242 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c5d67c78c-vvvv9" event={"ID":"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b","Type":"ContainerStarted","Data":"1b76f896485439da52fdf29afd297232a2c0e68cc26797c612e074e821199f04"} Nov 24 07:12:13 crc kubenswrapper[4809]: I1124 07:12:13.588280 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c5d67c78c-vvvv9" event={"ID":"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b","Type":"ContainerStarted","Data":"836551da23b0f009c6b905874fb4dcc6991e4507fc943fff6689538008022557"} Nov 24 07:12:13 crc kubenswrapper[4809]: I1124 07:12:13.588293 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c5d67c78c-vvvv9" event={"ID":"9c53f941-5f7d-489c-85eb-d07ff8d4ef6b","Type":"ContainerStarted","Data":"4dc4eb1a6814b28571e5a22ba71d5f19865864e4bf859091f0eef8d4eb098596"} Nov 24 07:12:13 crc kubenswrapper[4809]: I1124 07:12:13.588309 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:13 crc kubenswrapper[4809]: I1124 07:12:13.614556 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-ncjd6" podStartSLOduration=4.405515397 podStartE2EDuration="38.614536444s" podCreationTimestamp="2025-11-24 07:11:35 +0000 UTC" firstStartedPulling="2025-11-24 07:11:38.412725946 +0000 UTC m=+998.313317651" lastFinishedPulling="2025-11-24 07:12:12.621746973 +0000 UTC m=+1032.522338698" observedRunningTime="2025-11-24 07:12:13.605780582 +0000 UTC m=+1033.506372307" watchObservedRunningTime="2025-11-24 07:12:13.614536444 +0000 UTC m=+1033.515128149" Nov 24 07:12:13 crc kubenswrapper[4809]: I1124 07:12:13.653697 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6c5d67c78c-vvvv9" podStartSLOduration=4.653676172 podStartE2EDuration="4.653676172s" podCreationTimestamp="2025-11-24 07:12:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:12:13.636088155 +0000 UTC m=+1033.536679880" watchObservedRunningTime="2025-11-24 07:12:13.653676172 +0000 UTC m=+1033.554267877" Nov 24 07:12:14 crc kubenswrapper[4809]: I1124 07:12:14.685486 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 07:12:14 crc kubenswrapper[4809]: I1124 07:12:14.686233 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 07:12:14 crc kubenswrapper[4809]: I1124 07:12:14.720264 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 07:12:14 crc kubenswrapper[4809]: I1124 07:12:14.738057 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 07:12:15 crc kubenswrapper[4809]: I1124 07:12:15.161681 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:12:15 crc kubenswrapper[4809]: I1124 07:12:15.161751 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:12:15 crc kubenswrapper[4809]: I1124 07:12:15.298560 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:12:15 crc kubenswrapper[4809]: I1124 07:12:15.298618 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:12:15 crc kubenswrapper[4809]: I1124 07:12:15.611447 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 07:12:15 crc kubenswrapper[4809]: I1124 07:12:15.611525 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 07:12:15 crc kubenswrapper[4809]: I1124 07:12:15.777938 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 07:12:15 crc kubenswrapper[4809]: I1124 07:12:15.778959 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 07:12:15 crc kubenswrapper[4809]: I1124 07:12:15.837058 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 07:12:15 crc kubenswrapper[4809]: I1124 07:12:15.839346 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 07:12:16 crc kubenswrapper[4809]: I1124 07:12:16.619940 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 07:12:16 crc kubenswrapper[4809]: I1124 07:12:16.620000 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 07:12:17 crc kubenswrapper[4809]: I1124 07:12:17.601515 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 07:12:17 crc kubenswrapper[4809]: I1124 07:12:17.605440 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 07:12:17 crc kubenswrapper[4809]: I1124 07:12:17.645196 4809 generic.go:334] "Generic (PLEG): container finished" podID="6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293" containerID="0d7c52cf5c5cc03d5a34b01cf1a3e2ef1ec36b4324033c502207eeed4da969fa" exitCode=0 Nov 24 07:12:17 crc kubenswrapper[4809]: I1124 07:12:17.645438 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ncjd6" event={"ID":"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293","Type":"ContainerDied","Data":"0d7c52cf5c5cc03d5a34b01cf1a3e2ef1ec36b4324033c502207eeed4da969fa"} Nov 24 07:12:17 crc kubenswrapper[4809]: I1124 07:12:17.933152 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.033475 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-pvgsp"] Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.033712 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" podUID="946235c6-45ea-4bfe-98c1-24a8095b1843" containerName="dnsmasq-dns" containerID="cri-o://cb2b20004eca9a344d43c7fccb0cdf95cbe2ef16a181f66d0db8c1eaeca59b56" gracePeriod=10 Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.127758 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qsvl5" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.255478 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-credential-keys\") pod \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.255521 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4w9df\" (UniqueName: \"kubernetes.io/projected/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-kube-api-access-4w9df\") pod \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.255570 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-scripts\") pod \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.255673 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-config-data\") pod \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.255723 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-fernet-keys\") pod \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.255788 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-combined-ca-bundle\") pod \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\" (UID: \"f5d93390-ea3b-43d3-83ba-4d85fb8544d4\") " Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.263155 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "f5d93390-ea3b-43d3-83ba-4d85fb8544d4" (UID: "f5d93390-ea3b-43d3-83ba-4d85fb8544d4"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.266365 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f5d93390-ea3b-43d3-83ba-4d85fb8544d4" (UID: "f5d93390-ea3b-43d3-83ba-4d85fb8544d4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.266446 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-scripts" (OuterVolumeSpecName: "scripts") pod "f5d93390-ea3b-43d3-83ba-4d85fb8544d4" (UID: "f5d93390-ea3b-43d3-83ba-4d85fb8544d4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.268162 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-kube-api-access-4w9df" (OuterVolumeSpecName: "kube-api-access-4w9df") pod "f5d93390-ea3b-43d3-83ba-4d85fb8544d4" (UID: "f5d93390-ea3b-43d3-83ba-4d85fb8544d4"). InnerVolumeSpecName "kube-api-access-4w9df". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.290133 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f5d93390-ea3b-43d3-83ba-4d85fb8544d4" (UID: "f5d93390-ea3b-43d3-83ba-4d85fb8544d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.293109 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-config-data" (OuterVolumeSpecName: "config-data") pod "f5d93390-ea3b-43d3-83ba-4d85fb8544d4" (UID: "f5d93390-ea3b-43d3-83ba-4d85fb8544d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.359979 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.360010 4809 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.360018 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.360029 4809 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.360040 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4w9df\" (UniqueName: \"kubernetes.io/projected/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-kube-api-access-4w9df\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.360049 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5d93390-ea3b-43d3-83ba-4d85fb8544d4-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.489548 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.564531 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-ovsdbserver-sb\") pod \"946235c6-45ea-4bfe-98c1-24a8095b1843\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.564640 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6sz4\" (UniqueName: \"kubernetes.io/projected/946235c6-45ea-4bfe-98c1-24a8095b1843-kube-api-access-h6sz4\") pod \"946235c6-45ea-4bfe-98c1-24a8095b1843\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.564807 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-ovsdbserver-nb\") pod \"946235c6-45ea-4bfe-98c1-24a8095b1843\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.564847 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-dns-svc\") pod \"946235c6-45ea-4bfe-98c1-24a8095b1843\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.564878 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-dns-swift-storage-0\") pod \"946235c6-45ea-4bfe-98c1-24a8095b1843\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.564919 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-config\") pod \"946235c6-45ea-4bfe-98c1-24a8095b1843\" (UID: \"946235c6-45ea-4bfe-98c1-24a8095b1843\") " Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.583145 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/946235c6-45ea-4bfe-98c1-24a8095b1843-kube-api-access-h6sz4" (OuterVolumeSpecName: "kube-api-access-h6sz4") pod "946235c6-45ea-4bfe-98c1-24a8095b1843" (UID: "946235c6-45ea-4bfe-98c1-24a8095b1843"). InnerVolumeSpecName "kube-api-access-h6sz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.611821 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "946235c6-45ea-4bfe-98c1-24a8095b1843" (UID: "946235c6-45ea-4bfe-98c1-24a8095b1843"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.616950 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "946235c6-45ea-4bfe-98c1-24a8095b1843" (UID: "946235c6-45ea-4bfe-98c1-24a8095b1843"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.619473 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "946235c6-45ea-4bfe-98c1-24a8095b1843" (UID: "946235c6-45ea-4bfe-98c1-24a8095b1843"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.628795 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "946235c6-45ea-4bfe-98c1-24a8095b1843" (UID: "946235c6-45ea-4bfe-98c1-24a8095b1843"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.641749 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-config" (OuterVolumeSpecName: "config") pod "946235c6-45ea-4bfe-98c1-24a8095b1843" (UID: "946235c6-45ea-4bfe-98c1-24a8095b1843"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.664708 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qsvl5" event={"ID":"f5d93390-ea3b-43d3-83ba-4d85fb8544d4","Type":"ContainerDied","Data":"ec7caacc262d210a11c5099e18223026361c4d4be972849ac7376348764700bf"} Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.664829 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec7caacc262d210a11c5099e18223026361c4d4be972849ac7376348764700bf" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.664914 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qsvl5" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.671659 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.671684 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.671694 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.671703 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.671712 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/946235c6-45ea-4bfe-98c1-24a8095b1843-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.671721 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6sz4\" (UniqueName: \"kubernetes.io/projected/946235c6-45ea-4bfe-98c1-24a8095b1843-kube-api-access-h6sz4\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.674673 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"655e6590-673e-4fde-a097-1173dfcef9e2","Type":"ContainerStarted","Data":"2755fbe8785a1bb670bc1c46dc98b73feee7d48fd6630181c7e1097962fe8d83"} Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.679538 4809 generic.go:334] "Generic (PLEG): container finished" podID="946235c6-45ea-4bfe-98c1-24a8095b1843" containerID="cb2b20004eca9a344d43c7fccb0cdf95cbe2ef16a181f66d0db8c1eaeca59b56" exitCode=0 Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.679816 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.680492 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" event={"ID":"946235c6-45ea-4bfe-98c1-24a8095b1843","Type":"ContainerDied","Data":"cb2b20004eca9a344d43c7fccb0cdf95cbe2ef16a181f66d0db8c1eaeca59b56"} Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.680526 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-pvgsp" event={"ID":"946235c6-45ea-4bfe-98c1-24a8095b1843","Type":"ContainerDied","Data":"b085f99ad2c9e6b59d590a563ebf6cd6582ed3f3fe42f1a6558bd628afe4fc13"} Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.680543 4809 scope.go:117] "RemoveContainer" containerID="cb2b20004eca9a344d43c7fccb0cdf95cbe2ef16a181f66d0db8c1eaeca59b56" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.724552 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-pvgsp"] Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.727625 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-pvgsp"] Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.763796 4809 scope.go:117] "RemoveContainer" containerID="53b48d5ec8d1e3104e2c4a8ee666ec65ede0579a328ca3d848707ff42f8bb98a" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.797762 4809 scope.go:117] "RemoveContainer" containerID="cb2b20004eca9a344d43c7fccb0cdf95cbe2ef16a181f66d0db8c1eaeca59b56" Nov 24 07:12:18 crc kubenswrapper[4809]: E1124 07:12:18.798271 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb2b20004eca9a344d43c7fccb0cdf95cbe2ef16a181f66d0db8c1eaeca59b56\": container with ID starting with cb2b20004eca9a344d43c7fccb0cdf95cbe2ef16a181f66d0db8c1eaeca59b56 not found: ID does not exist" containerID="cb2b20004eca9a344d43c7fccb0cdf95cbe2ef16a181f66d0db8c1eaeca59b56" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.798326 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb2b20004eca9a344d43c7fccb0cdf95cbe2ef16a181f66d0db8c1eaeca59b56"} err="failed to get container status \"cb2b20004eca9a344d43c7fccb0cdf95cbe2ef16a181f66d0db8c1eaeca59b56\": rpc error: code = NotFound desc = could not find container \"cb2b20004eca9a344d43c7fccb0cdf95cbe2ef16a181f66d0db8c1eaeca59b56\": container with ID starting with cb2b20004eca9a344d43c7fccb0cdf95cbe2ef16a181f66d0db8c1eaeca59b56 not found: ID does not exist" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.798355 4809 scope.go:117] "RemoveContainer" containerID="53b48d5ec8d1e3104e2c4a8ee666ec65ede0579a328ca3d848707ff42f8bb98a" Nov 24 07:12:18 crc kubenswrapper[4809]: E1124 07:12:18.798695 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53b48d5ec8d1e3104e2c4a8ee666ec65ede0579a328ca3d848707ff42f8bb98a\": container with ID starting with 53b48d5ec8d1e3104e2c4a8ee666ec65ede0579a328ca3d848707ff42f8bb98a not found: ID does not exist" containerID="53b48d5ec8d1e3104e2c4a8ee666ec65ede0579a328ca3d848707ff42f8bb98a" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.798729 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53b48d5ec8d1e3104e2c4a8ee666ec65ede0579a328ca3d848707ff42f8bb98a"} err="failed to get container status \"53b48d5ec8d1e3104e2c4a8ee666ec65ede0579a328ca3d848707ff42f8bb98a\": rpc error: code = NotFound desc = could not find container \"53b48d5ec8d1e3104e2c4a8ee666ec65ede0579a328ca3d848707ff42f8bb98a\": container with ID starting with 53b48d5ec8d1e3104e2c4a8ee666ec65ede0579a328ca3d848707ff42f8bb98a not found: ID does not exist" Nov 24 07:12:18 crc kubenswrapper[4809]: I1124 07:12:18.915175 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="946235c6-45ea-4bfe-98c1-24a8095b1843" path="/var/lib/kubelet/pods/946235c6-45ea-4bfe-98c1-24a8095b1843/volumes" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.064807 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ncjd6" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.186827 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.186945 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.187023 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-config-data\") pod \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\" (UID: \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\") " Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.187762 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-combined-ca-bundle\") pod \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\" (UID: \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\") " Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.188166 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-scripts\") pod \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\" (UID: \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\") " Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.188278 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7bg7\" (UniqueName: \"kubernetes.io/projected/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-kube-api-access-t7bg7\") pod \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\" (UID: \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\") " Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.188407 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-logs\") pod \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\" (UID: \"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293\") " Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.189461 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-logs" (OuterVolumeSpecName: "logs") pod "6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293" (UID: "6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.197076 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-kube-api-access-t7bg7" (OuterVolumeSpecName: "kube-api-access-t7bg7") pod "6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293" (UID: "6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293"). InnerVolumeSpecName "kube-api-access-t7bg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.219108 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-scripts" (OuterVolumeSpecName: "scripts") pod "6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293" (UID: "6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.225816 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-fd8844f9c-6jj95"] Nov 24 07:12:19 crc kubenswrapper[4809]: E1124 07:12:19.226164 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946235c6-45ea-4bfe-98c1-24a8095b1843" containerName="init" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.226176 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="946235c6-45ea-4bfe-98c1-24a8095b1843" containerName="init" Nov 24 07:12:19 crc kubenswrapper[4809]: E1124 07:12:19.226185 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5d93390-ea3b-43d3-83ba-4d85fb8544d4" containerName="keystone-bootstrap" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.226191 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5d93390-ea3b-43d3-83ba-4d85fb8544d4" containerName="keystone-bootstrap" Nov 24 07:12:19 crc kubenswrapper[4809]: E1124 07:12:19.226211 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293" containerName="placement-db-sync" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.226216 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293" containerName="placement-db-sync" Nov 24 07:12:19 crc kubenswrapper[4809]: E1124 07:12:19.226228 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946235c6-45ea-4bfe-98c1-24a8095b1843" containerName="dnsmasq-dns" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.226235 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="946235c6-45ea-4bfe-98c1-24a8095b1843" containerName="dnsmasq-dns" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.226387 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5d93390-ea3b-43d3-83ba-4d85fb8544d4" containerName="keystone-bootstrap" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.226400 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="946235c6-45ea-4bfe-98c1-24a8095b1843" containerName="dnsmasq-dns" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.226409 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293" containerName="placement-db-sync" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.226918 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.238427 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.238470 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293" (UID: "6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.239293 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fkz7g" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.239508 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.239641 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.239762 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.240027 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.249719 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-fd8844f9c-6jj95"] Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.275612 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-config-data" (OuterVolumeSpecName: "config-data") pod "6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293" (UID: "6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.292877 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w44tb\" (UniqueName: \"kubernetes.io/projected/1a3f2bae-ab41-461b-87bc-455f850805e6-kube-api-access-w44tb\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.293018 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a3f2bae-ab41-461b-87bc-455f850805e6-scripts\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.293055 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a3f2bae-ab41-461b-87bc-455f850805e6-public-tls-certs\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.293115 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1a3f2bae-ab41-461b-87bc-455f850805e6-credential-keys\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.293136 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a3f2bae-ab41-461b-87bc-455f850805e6-internal-tls-certs\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.293156 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a3f2bae-ab41-461b-87bc-455f850805e6-config-data\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.293197 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a3f2bae-ab41-461b-87bc-455f850805e6-combined-ca-bundle\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.293219 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1a3f2bae-ab41-461b-87bc-455f850805e6-fernet-keys\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.293279 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.293289 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.293299 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.293310 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.293318 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7bg7\" (UniqueName: \"kubernetes.io/projected/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293-kube-api-access-t7bg7\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.395617 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1a3f2bae-ab41-461b-87bc-455f850805e6-fernet-keys\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.395679 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w44tb\" (UniqueName: \"kubernetes.io/projected/1a3f2bae-ab41-461b-87bc-455f850805e6-kube-api-access-w44tb\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.395731 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a3f2bae-ab41-461b-87bc-455f850805e6-scripts\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.395764 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a3f2bae-ab41-461b-87bc-455f850805e6-public-tls-certs\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.395809 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1a3f2bae-ab41-461b-87bc-455f850805e6-credential-keys\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.395827 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a3f2bae-ab41-461b-87bc-455f850805e6-internal-tls-certs\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.395845 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a3f2bae-ab41-461b-87bc-455f850805e6-config-data\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.395878 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a3f2bae-ab41-461b-87bc-455f850805e6-combined-ca-bundle\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.401732 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a3f2bae-ab41-461b-87bc-455f850805e6-combined-ca-bundle\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.402783 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a3f2bae-ab41-461b-87bc-455f850805e6-config-data\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.404521 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a3f2bae-ab41-461b-87bc-455f850805e6-internal-tls-certs\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.405740 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a3f2bae-ab41-461b-87bc-455f850805e6-public-tls-certs\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.405841 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1a3f2bae-ab41-461b-87bc-455f850805e6-fernet-keys\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.408233 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a3f2bae-ab41-461b-87bc-455f850805e6-scripts\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.412658 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1a3f2bae-ab41-461b-87bc-455f850805e6-credential-keys\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.416535 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w44tb\" (UniqueName: \"kubernetes.io/projected/1a3f2bae-ab41-461b-87bc-455f850805e6-kube-api-access-w44tb\") pod \"keystone-fd8844f9c-6jj95\" (UID: \"1a3f2bae-ab41-461b-87bc-455f850805e6\") " pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.601620 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.714838 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ncjd6" event={"ID":"6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293","Type":"ContainerDied","Data":"205905a1eddb1ff10ccefeaf24de603b33d2aeab2b41f9058b5c7991d4c231b9"} Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.715108 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="205905a1eddb1ff10ccefeaf24de603b33d2aeab2b41f9058b5c7991d4c231b9" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.715191 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ncjd6" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.720196 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-lkj6z" event={"ID":"972cb0ce-ea08-408c-a10d-ba1d14d42568","Type":"ContainerStarted","Data":"75d77f4aa40945e661f871d73ee743fffc6fcfa385d15c745bba59849623f594"} Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.739289 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-lkj6z" podStartSLOduration=3.501459157 podStartE2EDuration="44.739269972s" podCreationTimestamp="2025-11-24 07:11:35 +0000 UTC" firstStartedPulling="2025-11-24 07:11:38.119732797 +0000 UTC m=+998.020324502" lastFinishedPulling="2025-11-24 07:12:19.357543612 +0000 UTC m=+1039.258135317" observedRunningTime="2025-11-24 07:12:19.735587574 +0000 UTC m=+1039.636179279" watchObservedRunningTime="2025-11-24 07:12:19.739269972 +0000 UTC m=+1039.639861677" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.889891 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-78fc658c7d-ntbgd"] Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.892275 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.898959 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.899035 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.899396 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.901111 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.902414 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-5xnm8" Nov 24 07:12:19 crc kubenswrapper[4809]: I1124 07:12:19.906140 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-78fc658c7d-ntbgd"] Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.010656 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/446a68c7-316d-43a1-af00-feba9b01e2dd-scripts\") pod \"placement-78fc658c7d-ntbgd\" (UID: \"446a68c7-316d-43a1-af00-feba9b01e2dd\") " pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.010829 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/446a68c7-316d-43a1-af00-feba9b01e2dd-public-tls-certs\") pod \"placement-78fc658c7d-ntbgd\" (UID: \"446a68c7-316d-43a1-af00-feba9b01e2dd\") " pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.010875 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgrvr\" (UniqueName: \"kubernetes.io/projected/446a68c7-316d-43a1-af00-feba9b01e2dd-kube-api-access-hgrvr\") pod \"placement-78fc658c7d-ntbgd\" (UID: \"446a68c7-316d-43a1-af00-feba9b01e2dd\") " pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.011065 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/446a68c7-316d-43a1-af00-feba9b01e2dd-config-data\") pod \"placement-78fc658c7d-ntbgd\" (UID: \"446a68c7-316d-43a1-af00-feba9b01e2dd\") " pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.011100 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/446a68c7-316d-43a1-af00-feba9b01e2dd-internal-tls-certs\") pod \"placement-78fc658c7d-ntbgd\" (UID: \"446a68c7-316d-43a1-af00-feba9b01e2dd\") " pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.011263 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/446a68c7-316d-43a1-af00-feba9b01e2dd-combined-ca-bundle\") pod \"placement-78fc658c7d-ntbgd\" (UID: \"446a68c7-316d-43a1-af00-feba9b01e2dd\") " pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.014414 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/446a68c7-316d-43a1-af00-feba9b01e2dd-logs\") pod \"placement-78fc658c7d-ntbgd\" (UID: \"446a68c7-316d-43a1-af00-feba9b01e2dd\") " pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.036138 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-fd8844f9c-6jj95"] Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.130849 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/446a68c7-316d-43a1-af00-feba9b01e2dd-config-data\") pod \"placement-78fc658c7d-ntbgd\" (UID: \"446a68c7-316d-43a1-af00-feba9b01e2dd\") " pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.130907 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/446a68c7-316d-43a1-af00-feba9b01e2dd-internal-tls-certs\") pod \"placement-78fc658c7d-ntbgd\" (UID: \"446a68c7-316d-43a1-af00-feba9b01e2dd\") " pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.130955 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/446a68c7-316d-43a1-af00-feba9b01e2dd-combined-ca-bundle\") pod \"placement-78fc658c7d-ntbgd\" (UID: \"446a68c7-316d-43a1-af00-feba9b01e2dd\") " pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.131032 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/446a68c7-316d-43a1-af00-feba9b01e2dd-logs\") pod \"placement-78fc658c7d-ntbgd\" (UID: \"446a68c7-316d-43a1-af00-feba9b01e2dd\") " pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.131067 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/446a68c7-316d-43a1-af00-feba9b01e2dd-scripts\") pod \"placement-78fc658c7d-ntbgd\" (UID: \"446a68c7-316d-43a1-af00-feba9b01e2dd\") " pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.131098 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/446a68c7-316d-43a1-af00-feba9b01e2dd-public-tls-certs\") pod \"placement-78fc658c7d-ntbgd\" (UID: \"446a68c7-316d-43a1-af00-feba9b01e2dd\") " pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.131116 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgrvr\" (UniqueName: \"kubernetes.io/projected/446a68c7-316d-43a1-af00-feba9b01e2dd-kube-api-access-hgrvr\") pod \"placement-78fc658c7d-ntbgd\" (UID: \"446a68c7-316d-43a1-af00-feba9b01e2dd\") " pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.131761 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/446a68c7-316d-43a1-af00-feba9b01e2dd-logs\") pod \"placement-78fc658c7d-ntbgd\" (UID: \"446a68c7-316d-43a1-af00-feba9b01e2dd\") " pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.136682 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/446a68c7-316d-43a1-af00-feba9b01e2dd-internal-tls-certs\") pod \"placement-78fc658c7d-ntbgd\" (UID: \"446a68c7-316d-43a1-af00-feba9b01e2dd\") " pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.138402 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/446a68c7-316d-43a1-af00-feba9b01e2dd-scripts\") pod \"placement-78fc658c7d-ntbgd\" (UID: \"446a68c7-316d-43a1-af00-feba9b01e2dd\") " pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.141277 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/446a68c7-316d-43a1-af00-feba9b01e2dd-combined-ca-bundle\") pod \"placement-78fc658c7d-ntbgd\" (UID: \"446a68c7-316d-43a1-af00-feba9b01e2dd\") " pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.141600 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/446a68c7-316d-43a1-af00-feba9b01e2dd-public-tls-certs\") pod \"placement-78fc658c7d-ntbgd\" (UID: \"446a68c7-316d-43a1-af00-feba9b01e2dd\") " pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.145665 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/446a68c7-316d-43a1-af00-feba9b01e2dd-config-data\") pod \"placement-78fc658c7d-ntbgd\" (UID: \"446a68c7-316d-43a1-af00-feba9b01e2dd\") " pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.150566 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgrvr\" (UniqueName: \"kubernetes.io/projected/446a68c7-316d-43a1-af00-feba9b01e2dd-kube-api-access-hgrvr\") pod \"placement-78fc658c7d-ntbgd\" (UID: \"446a68c7-316d-43a1-af00-feba9b01e2dd\") " pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.232494 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.238927 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.792944 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-fd8844f9c-6jj95" event={"ID":"1a3f2bae-ab41-461b-87bc-455f850805e6","Type":"ContainerStarted","Data":"0ff8d79508fdbb32fefe9d4bfd5b06a022f47611fd6bede0dc6dd97e78e8f4e8"} Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.793287 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-fd8844f9c-6jj95" event={"ID":"1a3f2bae-ab41-461b-87bc-455f850805e6","Type":"ContainerStarted","Data":"b8e8b66980d15dbb946aeeca6554eb615a0c5624e8c21ce745d80736b40d9aa3"} Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.793305 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.801278 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-78fc658c7d-ntbgd"] Nov 24 07:12:20 crc kubenswrapper[4809]: I1124 07:12:20.834847 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-fd8844f9c-6jj95" podStartSLOduration=1.834828417 podStartE2EDuration="1.834828417s" podCreationTimestamp="2025-11-24 07:12:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:12:20.824286167 +0000 UTC m=+1040.724877872" watchObservedRunningTime="2025-11-24 07:12:20.834828417 +0000 UTC m=+1040.735420122" Nov 24 07:12:21 crc kubenswrapper[4809]: I1124 07:12:21.815355 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-78fc658c7d-ntbgd" event={"ID":"446a68c7-316d-43a1-af00-feba9b01e2dd","Type":"ContainerStarted","Data":"74ee70b8ce8a19be7f6875e09bb405bc2133423550da67aaf9631a09af69b87e"} Nov 24 07:12:21 crc kubenswrapper[4809]: I1124 07:12:21.816000 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-78fc658c7d-ntbgd" event={"ID":"446a68c7-316d-43a1-af00-feba9b01e2dd","Type":"ContainerStarted","Data":"a8de61f5539271b5f7023a217e35669463f3410756443880bb5dd4fba7d17fa4"} Nov 24 07:12:21 crc kubenswrapper[4809]: I1124 07:12:21.816015 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-78fc658c7d-ntbgd" event={"ID":"446a68c7-316d-43a1-af00-feba9b01e2dd","Type":"ContainerStarted","Data":"b270ad2cc5e369d495813513518f87eff78c7161817a75b9040f477d9f976268"} Nov 24 07:12:21 crc kubenswrapper[4809]: I1124 07:12:21.816182 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:21 crc kubenswrapper[4809]: I1124 07:12:21.818072 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-dtg82" event={"ID":"35dbbf41-295e-4ec9-bd02-12911c093ce8","Type":"ContainerStarted","Data":"c0a14a8affcc77fa698259e63f590d8e16e0632574118a5aaa166805e9e9aeed"} Nov 24 07:12:21 crc kubenswrapper[4809]: I1124 07:12:21.845616 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-78fc658c7d-ntbgd" podStartSLOduration=2.845581094 podStartE2EDuration="2.845581094s" podCreationTimestamp="2025-11-24 07:12:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:12:21.836398331 +0000 UTC m=+1041.736990106" watchObservedRunningTime="2025-11-24 07:12:21.845581094 +0000 UTC m=+1041.746172799" Nov 24 07:12:21 crc kubenswrapper[4809]: I1124 07:12:21.859238 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-dtg82" podStartSLOduration=4.58823054 podStartE2EDuration="46.859220556s" podCreationTimestamp="2025-11-24 07:11:35 +0000 UTC" firstStartedPulling="2025-11-24 07:11:38.106112896 +0000 UTC m=+998.006704601" lastFinishedPulling="2025-11-24 07:12:20.377102922 +0000 UTC m=+1040.277694617" observedRunningTime="2025-11-24 07:12:21.85033486 +0000 UTC m=+1041.750926565" watchObservedRunningTime="2025-11-24 07:12:21.859220556 +0000 UTC m=+1041.759812261" Nov 24 07:12:22 crc kubenswrapper[4809]: I1124 07:12:22.825701 4809 generic.go:334] "Generic (PLEG): container finished" podID="972cb0ce-ea08-408c-a10d-ba1d14d42568" containerID="75d77f4aa40945e661f871d73ee743fffc6fcfa385d15c745bba59849623f594" exitCode=0 Nov 24 07:12:22 crc kubenswrapper[4809]: I1124 07:12:22.825783 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-lkj6z" event={"ID":"972cb0ce-ea08-408c-a10d-ba1d14d42568","Type":"ContainerDied","Data":"75d77f4aa40945e661f871d73ee743fffc6fcfa385d15c745bba59849623f594"} Nov 24 07:12:22 crc kubenswrapper[4809]: I1124 07:12:22.826360 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:25 crc kubenswrapper[4809]: I1124 07:12:25.163451 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-79cbd445d4-5hkc8" podUID="5b3e7ca6-2814-4a14-b1f4-765f8086927e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Nov 24 07:12:25 crc kubenswrapper[4809]: I1124 07:12:25.304493 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-8585547888-qdlwv" podUID="b5487494-236a-4335-977f-11167eaa3cdc" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Nov 24 07:12:25 crc kubenswrapper[4809]: I1124 07:12:25.865902 4809 generic.go:334] "Generic (PLEG): container finished" podID="35dbbf41-295e-4ec9-bd02-12911c093ce8" containerID="c0a14a8affcc77fa698259e63f590d8e16e0632574118a5aaa166805e9e9aeed" exitCode=0 Nov 24 07:12:25 crc kubenswrapper[4809]: I1124 07:12:25.866101 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-dtg82" event={"ID":"35dbbf41-295e-4ec9-bd02-12911c093ce8","Type":"ContainerDied","Data":"c0a14a8affcc77fa698259e63f590d8e16e0632574118a5aaa166805e9e9aeed"} Nov 24 07:12:26 crc kubenswrapper[4809]: I1124 07:12:26.671403 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-lkj6z" Nov 24 07:12:26 crc kubenswrapper[4809]: I1124 07:12:26.770609 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/972cb0ce-ea08-408c-a10d-ba1d14d42568-combined-ca-bundle\") pod \"972cb0ce-ea08-408c-a10d-ba1d14d42568\" (UID: \"972cb0ce-ea08-408c-a10d-ba1d14d42568\") " Nov 24 07:12:26 crc kubenswrapper[4809]: I1124 07:12:26.770688 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lf79z\" (UniqueName: \"kubernetes.io/projected/972cb0ce-ea08-408c-a10d-ba1d14d42568-kube-api-access-lf79z\") pod \"972cb0ce-ea08-408c-a10d-ba1d14d42568\" (UID: \"972cb0ce-ea08-408c-a10d-ba1d14d42568\") " Nov 24 07:12:26 crc kubenswrapper[4809]: I1124 07:12:26.770750 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/972cb0ce-ea08-408c-a10d-ba1d14d42568-db-sync-config-data\") pod \"972cb0ce-ea08-408c-a10d-ba1d14d42568\" (UID: \"972cb0ce-ea08-408c-a10d-ba1d14d42568\") " Nov 24 07:12:26 crc kubenswrapper[4809]: I1124 07:12:26.775693 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/972cb0ce-ea08-408c-a10d-ba1d14d42568-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "972cb0ce-ea08-408c-a10d-ba1d14d42568" (UID: "972cb0ce-ea08-408c-a10d-ba1d14d42568"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:26 crc kubenswrapper[4809]: I1124 07:12:26.778237 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/972cb0ce-ea08-408c-a10d-ba1d14d42568-kube-api-access-lf79z" (OuterVolumeSpecName: "kube-api-access-lf79z") pod "972cb0ce-ea08-408c-a10d-ba1d14d42568" (UID: "972cb0ce-ea08-408c-a10d-ba1d14d42568"). InnerVolumeSpecName "kube-api-access-lf79z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:12:26 crc kubenswrapper[4809]: I1124 07:12:26.806269 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/972cb0ce-ea08-408c-a10d-ba1d14d42568-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "972cb0ce-ea08-408c-a10d-ba1d14d42568" (UID: "972cb0ce-ea08-408c-a10d-ba1d14d42568"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:26 crc kubenswrapper[4809]: I1124 07:12:26.873508 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/972cb0ce-ea08-408c-a10d-ba1d14d42568-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:26 crc kubenswrapper[4809]: I1124 07:12:26.873727 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lf79z\" (UniqueName: \"kubernetes.io/projected/972cb0ce-ea08-408c-a10d-ba1d14d42568-kube-api-access-lf79z\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:26 crc kubenswrapper[4809]: I1124 07:12:26.873806 4809 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/972cb0ce-ea08-408c-a10d-ba1d14d42568-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:26 crc kubenswrapper[4809]: I1124 07:12:26.877816 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-lkj6z" Nov 24 07:12:26 crc kubenswrapper[4809]: I1124 07:12:26.877958 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-lkj6z" event={"ID":"972cb0ce-ea08-408c-a10d-ba1d14d42568","Type":"ContainerDied","Data":"d8c83e0e63e7f83fcb57551ac76044b91d9fd7031f91729f158ae5ba92506e3e"} Nov 24 07:12:26 crc kubenswrapper[4809]: I1124 07:12:26.878028 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8c83e0e63e7f83fcb57551ac76044b91d9fd7031f91729f158ae5ba92506e3e" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.186359 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-dtg82" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.279797 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-scripts\") pod \"35dbbf41-295e-4ec9-bd02-12911c093ce8\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.279848 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9q8n6\" (UniqueName: \"kubernetes.io/projected/35dbbf41-295e-4ec9-bd02-12911c093ce8-kube-api-access-9q8n6\") pod \"35dbbf41-295e-4ec9-bd02-12911c093ce8\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.280030 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-config-data\") pod \"35dbbf41-295e-4ec9-bd02-12911c093ce8\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.280100 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/35dbbf41-295e-4ec9-bd02-12911c093ce8-etc-machine-id\") pod \"35dbbf41-295e-4ec9-bd02-12911c093ce8\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.280128 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-db-sync-config-data\") pod \"35dbbf41-295e-4ec9-bd02-12911c093ce8\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.280147 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/35dbbf41-295e-4ec9-bd02-12911c093ce8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "35dbbf41-295e-4ec9-bd02-12911c093ce8" (UID: "35dbbf41-295e-4ec9-bd02-12911c093ce8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.280191 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-combined-ca-bundle\") pod \"35dbbf41-295e-4ec9-bd02-12911c093ce8\" (UID: \"35dbbf41-295e-4ec9-bd02-12911c093ce8\") " Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.280595 4809 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/35dbbf41-295e-4ec9-bd02-12911c093ce8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.283942 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35dbbf41-295e-4ec9-bd02-12911c093ce8-kube-api-access-9q8n6" (OuterVolumeSpecName: "kube-api-access-9q8n6") pod "35dbbf41-295e-4ec9-bd02-12911c093ce8" (UID: "35dbbf41-295e-4ec9-bd02-12911c093ce8"). InnerVolumeSpecName "kube-api-access-9q8n6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.284929 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-scripts" (OuterVolumeSpecName: "scripts") pod "35dbbf41-295e-4ec9-bd02-12911c093ce8" (UID: "35dbbf41-295e-4ec9-bd02-12911c093ce8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.287035 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "35dbbf41-295e-4ec9-bd02-12911c093ce8" (UID: "35dbbf41-295e-4ec9-bd02-12911c093ce8"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.306793 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35dbbf41-295e-4ec9-bd02-12911c093ce8" (UID: "35dbbf41-295e-4ec9-bd02-12911c093ce8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.326714 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-config-data" (OuterVolumeSpecName: "config-data") pod "35dbbf41-295e-4ec9-bd02-12911c093ce8" (UID: "35dbbf41-295e-4ec9-bd02-12911c093ce8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.381945 4809 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.382030 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.382042 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.382051 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9q8n6\" (UniqueName: \"kubernetes.io/projected/35dbbf41-295e-4ec9-bd02-12911c093ce8-kube-api-access-9q8n6\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.382061 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35dbbf41-295e-4ec9-bd02-12911c093ce8-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.959254 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="655e6590-673e-4fde-a097-1173dfcef9e2" containerName="ceilometer-central-agent" containerID="cri-o://42f9be194d69131769e788b4521d542fb8b90d88ac5fe9d2fc53be376177c837" gracePeriod=30 Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.959611 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="655e6590-673e-4fde-a097-1173dfcef9e2" containerName="proxy-httpd" containerID="cri-o://a64a2f12bea992c607246a7a39249b1ccb7b9647115c0abd676355a3ddaebdb3" gracePeriod=30 Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.959672 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="655e6590-673e-4fde-a097-1173dfcef9e2" containerName="sg-core" containerID="cri-o://2755fbe8785a1bb670bc1c46dc98b73feee7d48fd6630181c7e1097962fe8d83" gracePeriod=30 Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.959720 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="655e6590-673e-4fde-a097-1173dfcef9e2" containerName="ceilometer-notification-agent" containerID="cri-o://ffc5ddd380aa10bc65340c8e0facbe2c23a4aba66abadd3808bb97c90888fe85" gracePeriod=30 Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.959491 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"655e6590-673e-4fde-a097-1173dfcef9e2","Type":"ContainerStarted","Data":"a64a2f12bea992c607246a7a39249b1ccb7b9647115c0abd676355a3ddaebdb3"} Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.959782 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.981318 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-dtg82" event={"ID":"35dbbf41-295e-4ec9-bd02-12911c093ce8","Type":"ContainerDied","Data":"54e9341ce2cb545e234b9294c9745fb3e37ad1a25f0074ca172c45b53ec66ef3"} Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.981555 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54e9341ce2cb545e234b9294c9745fb3e37ad1a25f0074ca172c45b53ec66ef3" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.981665 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-dtg82" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.983592 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7947c766c8-w6s92"] Nov 24 07:12:27 crc kubenswrapper[4809]: E1124 07:12:27.984073 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35dbbf41-295e-4ec9-bd02-12911c093ce8" containerName="cinder-db-sync" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.984091 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="35dbbf41-295e-4ec9-bd02-12911c093ce8" containerName="cinder-db-sync" Nov 24 07:12:27 crc kubenswrapper[4809]: E1124 07:12:27.984115 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="972cb0ce-ea08-408c-a10d-ba1d14d42568" containerName="barbican-db-sync" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.984123 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="972cb0ce-ea08-408c-a10d-ba1d14d42568" containerName="barbican-db-sync" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.984297 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="35dbbf41-295e-4ec9-bd02-12911c093ce8" containerName="cinder-db-sync" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.984326 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="972cb0ce-ea08-408c-a10d-ba1d14d42568" containerName="barbican-db-sync" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.985320 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7947c766c8-w6s92" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.988658 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.988880 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-dvvzk" Nov 24 07:12:27 crc kubenswrapper[4809]: I1124 07:12:27.992353 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.024869 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-cc9bfd6b9-27jr8"] Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.026729 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-cc9bfd6b9-27jr8" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.030339 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.057211 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7947c766c8-w6s92"] Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.084021 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-cc9bfd6b9-27jr8"] Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.085368 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.582575008 podStartE2EDuration="53.085358672s" podCreationTimestamp="2025-11-24 07:11:35 +0000 UTC" firstStartedPulling="2025-11-24 07:11:37.174315113 +0000 UTC m=+997.074906818" lastFinishedPulling="2025-11-24 07:12:26.677098777 +0000 UTC m=+1046.577690482" observedRunningTime="2025-11-24 07:12:28.013361644 +0000 UTC m=+1047.913953349" watchObservedRunningTime="2025-11-24 07:12:28.085358672 +0000 UTC m=+1047.985950377" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.097854 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwpjc\" (UniqueName: \"kubernetes.io/projected/ada052fc-743e-477f-b5be-848ca3536f09-kube-api-access-nwpjc\") pod \"barbican-keystone-listener-7947c766c8-w6s92\" (UID: \"ada052fc-743e-477f-b5be-848ca3536f09\") " pod="openstack/barbican-keystone-listener-7947c766c8-w6s92" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.097914 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c05e3fc-9407-4bc4-86b4-409ebee7f3e7-config-data\") pod \"barbican-worker-cc9bfd6b9-27jr8\" (UID: \"5c05e3fc-9407-4bc4-86b4-409ebee7f3e7\") " pod="openstack/barbican-worker-cc9bfd6b9-27jr8" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.097981 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c05e3fc-9407-4bc4-86b4-409ebee7f3e7-combined-ca-bundle\") pod \"barbican-worker-cc9bfd6b9-27jr8\" (UID: \"5c05e3fc-9407-4bc4-86b4-409ebee7f3e7\") " pod="openstack/barbican-worker-cc9bfd6b9-27jr8" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.098011 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72x9x\" (UniqueName: \"kubernetes.io/projected/5c05e3fc-9407-4bc4-86b4-409ebee7f3e7-kube-api-access-72x9x\") pod \"barbican-worker-cc9bfd6b9-27jr8\" (UID: \"5c05e3fc-9407-4bc4-86b4-409ebee7f3e7\") " pod="openstack/barbican-worker-cc9bfd6b9-27jr8" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.098059 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ada052fc-743e-477f-b5be-848ca3536f09-combined-ca-bundle\") pod \"barbican-keystone-listener-7947c766c8-w6s92\" (UID: \"ada052fc-743e-477f-b5be-848ca3536f09\") " pod="openstack/barbican-keystone-listener-7947c766c8-w6s92" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.098076 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c05e3fc-9407-4bc4-86b4-409ebee7f3e7-logs\") pod \"barbican-worker-cc9bfd6b9-27jr8\" (UID: \"5c05e3fc-9407-4bc4-86b4-409ebee7f3e7\") " pod="openstack/barbican-worker-cc9bfd6b9-27jr8" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.098098 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ada052fc-743e-477f-b5be-848ca3536f09-config-data\") pod \"barbican-keystone-listener-7947c766c8-w6s92\" (UID: \"ada052fc-743e-477f-b5be-848ca3536f09\") " pod="openstack/barbican-keystone-listener-7947c766c8-w6s92" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.099929 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ada052fc-743e-477f-b5be-848ca3536f09-config-data-custom\") pod \"barbican-keystone-listener-7947c766c8-w6s92\" (UID: \"ada052fc-743e-477f-b5be-848ca3536f09\") " pod="openstack/barbican-keystone-listener-7947c766c8-w6s92" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.100223 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c05e3fc-9407-4bc4-86b4-409ebee7f3e7-config-data-custom\") pod \"barbican-worker-cc9bfd6b9-27jr8\" (UID: \"5c05e3fc-9407-4bc4-86b4-409ebee7f3e7\") " pod="openstack/barbican-worker-cc9bfd6b9-27jr8" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.100260 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ada052fc-743e-477f-b5be-848ca3536f09-logs\") pod \"barbican-keystone-listener-7947c766c8-w6s92\" (UID: \"ada052fc-743e-477f-b5be-848ca3536f09\") " pod="openstack/barbican-keystone-listener-7947c766c8-w6s92" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.128410 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-xrnts"] Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.130041 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-xrnts" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.158477 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-xrnts"] Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.201396 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c05e3fc-9407-4bc4-86b4-409ebee7f3e7-config-data-custom\") pod \"barbican-worker-cc9bfd6b9-27jr8\" (UID: \"5c05e3fc-9407-4bc4-86b4-409ebee7f3e7\") " pod="openstack/barbican-worker-cc9bfd6b9-27jr8" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.201438 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ada052fc-743e-477f-b5be-848ca3536f09-logs\") pod \"barbican-keystone-listener-7947c766c8-w6s92\" (UID: \"ada052fc-743e-477f-b5be-848ca3536f09\") " pod="openstack/barbican-keystone-listener-7947c766c8-w6s92" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.201474 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwpjc\" (UniqueName: \"kubernetes.io/projected/ada052fc-743e-477f-b5be-848ca3536f09-kube-api-access-nwpjc\") pod \"barbican-keystone-listener-7947c766c8-w6s92\" (UID: \"ada052fc-743e-477f-b5be-848ca3536f09\") " pod="openstack/barbican-keystone-listener-7947c766c8-w6s92" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.201497 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-xrnts\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " pod="openstack/dnsmasq-dns-85ff748b95-xrnts" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.201519 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmf5f\" (UniqueName: \"kubernetes.io/projected/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-kube-api-access-pmf5f\") pod \"dnsmasq-dns-85ff748b95-xrnts\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " pod="openstack/dnsmasq-dns-85ff748b95-xrnts" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.201538 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-xrnts\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " pod="openstack/dnsmasq-dns-85ff748b95-xrnts" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.201559 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c05e3fc-9407-4bc4-86b4-409ebee7f3e7-config-data\") pod \"barbican-worker-cc9bfd6b9-27jr8\" (UID: \"5c05e3fc-9407-4bc4-86b4-409ebee7f3e7\") " pod="openstack/barbican-worker-cc9bfd6b9-27jr8" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.201592 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-dns-svc\") pod \"dnsmasq-dns-85ff748b95-xrnts\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " pod="openstack/dnsmasq-dns-85ff748b95-xrnts" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.201610 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-xrnts\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " pod="openstack/dnsmasq-dns-85ff748b95-xrnts" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.201637 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c05e3fc-9407-4bc4-86b4-409ebee7f3e7-combined-ca-bundle\") pod \"barbican-worker-cc9bfd6b9-27jr8\" (UID: \"5c05e3fc-9407-4bc4-86b4-409ebee7f3e7\") " pod="openstack/barbican-worker-cc9bfd6b9-27jr8" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.201674 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72x9x\" (UniqueName: \"kubernetes.io/projected/5c05e3fc-9407-4bc4-86b4-409ebee7f3e7-kube-api-access-72x9x\") pod \"barbican-worker-cc9bfd6b9-27jr8\" (UID: \"5c05e3fc-9407-4bc4-86b4-409ebee7f3e7\") " pod="openstack/barbican-worker-cc9bfd6b9-27jr8" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.201712 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ada052fc-743e-477f-b5be-848ca3536f09-combined-ca-bundle\") pod \"barbican-keystone-listener-7947c766c8-w6s92\" (UID: \"ada052fc-743e-477f-b5be-848ca3536f09\") " pod="openstack/barbican-keystone-listener-7947c766c8-w6s92" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.201738 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c05e3fc-9407-4bc4-86b4-409ebee7f3e7-logs\") pod \"barbican-worker-cc9bfd6b9-27jr8\" (UID: \"5c05e3fc-9407-4bc4-86b4-409ebee7f3e7\") " pod="openstack/barbican-worker-cc9bfd6b9-27jr8" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.201754 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-config\") pod \"dnsmasq-dns-85ff748b95-xrnts\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " pod="openstack/dnsmasq-dns-85ff748b95-xrnts" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.201775 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ada052fc-743e-477f-b5be-848ca3536f09-config-data\") pod \"barbican-keystone-listener-7947c766c8-w6s92\" (UID: \"ada052fc-743e-477f-b5be-848ca3536f09\") " pod="openstack/barbican-keystone-listener-7947c766c8-w6s92" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.201791 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ada052fc-743e-477f-b5be-848ca3536f09-config-data-custom\") pod \"barbican-keystone-listener-7947c766c8-w6s92\" (UID: \"ada052fc-743e-477f-b5be-848ca3536f09\") " pod="openstack/barbican-keystone-listener-7947c766c8-w6s92" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.203601 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ada052fc-743e-477f-b5be-848ca3536f09-logs\") pod \"barbican-keystone-listener-7947c766c8-w6s92\" (UID: \"ada052fc-743e-477f-b5be-848ca3536f09\") " pod="openstack/barbican-keystone-listener-7947c766c8-w6s92" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.212567 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c05e3fc-9407-4bc4-86b4-409ebee7f3e7-logs\") pod \"barbican-worker-cc9bfd6b9-27jr8\" (UID: \"5c05e3fc-9407-4bc4-86b4-409ebee7f3e7\") " pod="openstack/barbican-worker-cc9bfd6b9-27jr8" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.221037 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ada052fc-743e-477f-b5be-848ca3536f09-config-data-custom\") pod \"barbican-keystone-listener-7947c766c8-w6s92\" (UID: \"ada052fc-743e-477f-b5be-848ca3536f09\") " pod="openstack/barbican-keystone-listener-7947c766c8-w6s92" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.221656 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c05e3fc-9407-4bc4-86b4-409ebee7f3e7-config-data-custom\") pod \"barbican-worker-cc9bfd6b9-27jr8\" (UID: \"5c05e3fc-9407-4bc4-86b4-409ebee7f3e7\") " pod="openstack/barbican-worker-cc9bfd6b9-27jr8" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.228658 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c05e3fc-9407-4bc4-86b4-409ebee7f3e7-combined-ca-bundle\") pod \"barbican-worker-cc9bfd6b9-27jr8\" (UID: \"5c05e3fc-9407-4bc4-86b4-409ebee7f3e7\") " pod="openstack/barbican-worker-cc9bfd6b9-27jr8" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.237509 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ada052fc-743e-477f-b5be-848ca3536f09-combined-ca-bundle\") pod \"barbican-keystone-listener-7947c766c8-w6s92\" (UID: \"ada052fc-743e-477f-b5be-848ca3536f09\") " pod="openstack/barbican-keystone-listener-7947c766c8-w6s92" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.239215 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwpjc\" (UniqueName: \"kubernetes.io/projected/ada052fc-743e-477f-b5be-848ca3536f09-kube-api-access-nwpjc\") pod \"barbican-keystone-listener-7947c766c8-w6s92\" (UID: \"ada052fc-743e-477f-b5be-848ca3536f09\") " pod="openstack/barbican-keystone-listener-7947c766c8-w6s92" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.242143 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c05e3fc-9407-4bc4-86b4-409ebee7f3e7-config-data\") pod \"barbican-worker-cc9bfd6b9-27jr8\" (UID: \"5c05e3fc-9407-4bc4-86b4-409ebee7f3e7\") " pod="openstack/barbican-worker-cc9bfd6b9-27jr8" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.247058 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ada052fc-743e-477f-b5be-848ca3536f09-config-data\") pod \"barbican-keystone-listener-7947c766c8-w6s92\" (UID: \"ada052fc-743e-477f-b5be-848ca3536f09\") " pod="openstack/barbican-keystone-listener-7947c766c8-w6s92" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.259007 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72x9x\" (UniqueName: \"kubernetes.io/projected/5c05e3fc-9407-4bc4-86b4-409ebee7f3e7-kube-api-access-72x9x\") pod \"barbican-worker-cc9bfd6b9-27jr8\" (UID: \"5c05e3fc-9407-4bc4-86b4-409ebee7f3e7\") " pod="openstack/barbican-worker-cc9bfd6b9-27jr8" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.259526 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6bd6d9b5b4-rvvrx"] Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.260855 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.266276 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.299987 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6bd6d9b5b4-rvvrx"] Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.305141 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-combined-ca-bundle\") pod \"barbican-api-6bd6d9b5b4-rvvrx\" (UID: \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\") " pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.305219 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-config\") pod \"dnsmasq-dns-85ff748b95-xrnts\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " pod="openstack/dnsmasq-dns-85ff748b95-xrnts" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.305246 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g58b\" (UniqueName: \"kubernetes.io/projected/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-kube-api-access-2g58b\") pod \"barbican-api-6bd6d9b5b4-rvvrx\" (UID: \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\") " pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.305268 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-config-data\") pod \"barbican-api-6bd6d9b5b4-rvvrx\" (UID: \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\") " pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.305289 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-config-data-custom\") pod \"barbican-api-6bd6d9b5b4-rvvrx\" (UID: \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\") " pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.305308 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-logs\") pod \"barbican-api-6bd6d9b5b4-rvvrx\" (UID: \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\") " pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.305355 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-xrnts\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " pod="openstack/dnsmasq-dns-85ff748b95-xrnts" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.305376 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmf5f\" (UniqueName: \"kubernetes.io/projected/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-kube-api-access-pmf5f\") pod \"dnsmasq-dns-85ff748b95-xrnts\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " pod="openstack/dnsmasq-dns-85ff748b95-xrnts" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.305395 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-xrnts\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " pod="openstack/dnsmasq-dns-85ff748b95-xrnts" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.305437 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-dns-svc\") pod \"dnsmasq-dns-85ff748b95-xrnts\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " pod="openstack/dnsmasq-dns-85ff748b95-xrnts" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.305457 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-xrnts\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " pod="openstack/dnsmasq-dns-85ff748b95-xrnts" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.306366 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-xrnts\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " pod="openstack/dnsmasq-dns-85ff748b95-xrnts" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.306877 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-config\") pod \"dnsmasq-dns-85ff748b95-xrnts\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " pod="openstack/dnsmasq-dns-85ff748b95-xrnts" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.307407 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-xrnts\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " pod="openstack/dnsmasq-dns-85ff748b95-xrnts" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.312779 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-dns-svc\") pod \"dnsmasq-dns-85ff748b95-xrnts\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " pod="openstack/dnsmasq-dns-85ff748b95-xrnts" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.313140 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7947c766c8-w6s92" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.317673 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-xrnts\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " pod="openstack/dnsmasq-dns-85ff748b95-xrnts" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.360205 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmf5f\" (UniqueName: \"kubernetes.io/projected/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-kube-api-access-pmf5f\") pod \"dnsmasq-dns-85ff748b95-xrnts\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " pod="openstack/dnsmasq-dns-85ff748b95-xrnts" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.390097 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.391613 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.398866 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.399719 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.399866 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-6m26k" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.400008 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.408959 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-combined-ca-bundle\") pod \"barbican-api-6bd6d9b5b4-rvvrx\" (UID: \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\") " pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.409031 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g58b\" (UniqueName: \"kubernetes.io/projected/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-kube-api-access-2g58b\") pod \"barbican-api-6bd6d9b5b4-rvvrx\" (UID: \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\") " pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.409055 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-config-data\") pod \"barbican-api-6bd6d9b5b4-rvvrx\" (UID: \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\") " pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.409073 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-config-data-custom\") pod \"barbican-api-6bd6d9b5b4-rvvrx\" (UID: \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\") " pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.409093 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-logs\") pod \"barbican-api-6bd6d9b5b4-rvvrx\" (UID: \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\") " pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.409582 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-logs\") pod \"barbican-api-6bd6d9b5b4-rvvrx\" (UID: \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\") " pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.419165 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-config-data\") pod \"barbican-api-6bd6d9b5b4-rvvrx\" (UID: \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\") " pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.419747 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-combined-ca-bundle\") pod \"barbican-api-6bd6d9b5b4-rvvrx\" (UID: \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\") " pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.441160 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g58b\" (UniqueName: \"kubernetes.io/projected/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-kube-api-access-2g58b\") pod \"barbican-api-6bd6d9b5b4-rvvrx\" (UID: \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\") " pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.448792 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-xrnts"] Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.449229 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-xrnts" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.453683 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-config-data-custom\") pod \"barbican-api-6bd6d9b5b4-rvvrx\" (UID: \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\") " pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.454098 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-cc9bfd6b9-27jr8" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.464439 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.471225 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-br547"] Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.472956 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.492862 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-br547"] Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.533239 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.536866 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-config-data\") pod \"cinder-scheduler-0\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.536947 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-br547\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.537155 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl4dp\" (UniqueName: \"kubernetes.io/projected/6315ca3f-8d9f-4e7e-b326-725025544aed-kube-api-access-fl4dp\") pod \"cinder-scheduler-0\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.537230 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-br547\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.538682 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-scripts\") pod \"cinder-scheduler-0\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.538733 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-config\") pod \"dnsmasq-dns-5c9776ccc5-br547\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.538764 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.538859 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtbh2\" (UniqueName: \"kubernetes.io/projected/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-kube-api-access-wtbh2\") pod \"dnsmasq-dns-5c9776ccc5-br547\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.538996 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-br547\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.539084 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6315ca3f-8d9f-4e7e-b326-725025544aed-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.539285 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-br547\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.620157 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.621626 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.626747 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.628416 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.652595 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtbh2\" (UniqueName: \"kubernetes.io/projected/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-kube-api-access-wtbh2\") pod \"dnsmasq-dns-5c9776ccc5-br547\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.652688 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-br547\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.652759 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6315ca3f-8d9f-4e7e-b326-725025544aed-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.652877 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-br547\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.653053 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.653090 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-config-data\") pod \"cinder-scheduler-0\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.653112 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-br547\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.653814 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl4dp\" (UniqueName: \"kubernetes.io/projected/6315ca3f-8d9f-4e7e-b326-725025544aed-kube-api-access-fl4dp\") pod \"cinder-scheduler-0\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.653839 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-br547\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.653913 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-scripts\") pod \"cinder-scheduler-0\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.653939 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-config\") pod \"dnsmasq-dns-5c9776ccc5-br547\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.653959 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.659678 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-br547\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.659730 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6315ca3f-8d9f-4e7e-b326-725025544aed-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.660369 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-br547\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.661858 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.662064 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-br547\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.663255 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-br547\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.665510 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-config\") pod \"dnsmasq-dns-5c9776ccc5-br547\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.665654 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.665753 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-scripts\") pod \"cinder-scheduler-0\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.668678 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-config-data\") pod \"cinder-scheduler-0\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.679507 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtbh2\" (UniqueName: \"kubernetes.io/projected/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-kube-api-access-wtbh2\") pod \"dnsmasq-dns-5c9776ccc5-br547\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.688773 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl4dp\" (UniqueName: \"kubernetes.io/projected/6315ca3f-8d9f-4e7e-b326-725025544aed-kube-api-access-fl4dp\") pod \"cinder-scheduler-0\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.722923 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.750524 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.762680 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-config-data\") pod \"cinder-api-0\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " pod="openstack/cinder-api-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.762894 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df25l\" (UniqueName: \"kubernetes.io/projected/1a14aa5c-2f38-4db8-9bf9-422c14699895-kube-api-access-df25l\") pod \"cinder-api-0\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " pod="openstack/cinder-api-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.763217 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-config-data-custom\") pod \"cinder-api-0\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " pod="openstack/cinder-api-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.763916 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-scripts\") pod \"cinder-api-0\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " pod="openstack/cinder-api-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.763986 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a14aa5c-2f38-4db8-9bf9-422c14699895-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " pod="openstack/cinder-api-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.764061 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a14aa5c-2f38-4db8-9bf9-422c14699895-logs\") pod \"cinder-api-0\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " pod="openstack/cinder-api-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.764126 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " pod="openstack/cinder-api-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.835163 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.867369 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-config-data\") pod \"cinder-api-0\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " pod="openstack/cinder-api-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.867477 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df25l\" (UniqueName: \"kubernetes.io/projected/1a14aa5c-2f38-4db8-9bf9-422c14699895-kube-api-access-df25l\") pod \"cinder-api-0\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " pod="openstack/cinder-api-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.867526 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-config-data-custom\") pod \"cinder-api-0\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " pod="openstack/cinder-api-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.867574 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-scripts\") pod \"cinder-api-0\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " pod="openstack/cinder-api-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.867595 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a14aa5c-2f38-4db8-9bf9-422c14699895-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " pod="openstack/cinder-api-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.867626 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a14aa5c-2f38-4db8-9bf9-422c14699895-logs\") pod \"cinder-api-0\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " pod="openstack/cinder-api-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.867675 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " pod="openstack/cinder-api-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.870701 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a14aa5c-2f38-4db8-9bf9-422c14699895-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " pod="openstack/cinder-api-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.871509 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a14aa5c-2f38-4db8-9bf9-422c14699895-logs\") pod \"cinder-api-0\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " pod="openstack/cinder-api-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.874545 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-config-data-custom\") pod \"cinder-api-0\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " pod="openstack/cinder-api-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.874994 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " pod="openstack/cinder-api-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.877988 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-config-data\") pod \"cinder-api-0\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " pod="openstack/cinder-api-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.879877 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-scripts\") pod \"cinder-api-0\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " pod="openstack/cinder-api-0" Nov 24 07:12:28 crc kubenswrapper[4809]: I1124 07:12:28.886585 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df25l\" (UniqueName: \"kubernetes.io/projected/1a14aa5c-2f38-4db8-9bf9-422c14699895-kube-api-access-df25l\") pod \"cinder-api-0\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " pod="openstack/cinder-api-0" Nov 24 07:12:29 crc kubenswrapper[4809]: I1124 07:12:29.023182 4809 generic.go:334] "Generic (PLEG): container finished" podID="655e6590-673e-4fde-a097-1173dfcef9e2" containerID="a64a2f12bea992c607246a7a39249b1ccb7b9647115c0abd676355a3ddaebdb3" exitCode=0 Nov 24 07:12:29 crc kubenswrapper[4809]: I1124 07:12:29.023218 4809 generic.go:334] "Generic (PLEG): container finished" podID="655e6590-673e-4fde-a097-1173dfcef9e2" containerID="2755fbe8785a1bb670bc1c46dc98b73feee7d48fd6630181c7e1097962fe8d83" exitCode=2 Nov 24 07:12:29 crc kubenswrapper[4809]: I1124 07:12:29.023228 4809 generic.go:334] "Generic (PLEG): container finished" podID="655e6590-673e-4fde-a097-1173dfcef9e2" containerID="42f9be194d69131769e788b4521d542fb8b90d88ac5fe9d2fc53be376177c837" exitCode=0 Nov 24 07:12:29 crc kubenswrapper[4809]: I1124 07:12:29.023251 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"655e6590-673e-4fde-a097-1173dfcef9e2","Type":"ContainerDied","Data":"a64a2f12bea992c607246a7a39249b1ccb7b9647115c0abd676355a3ddaebdb3"} Nov 24 07:12:29 crc kubenswrapper[4809]: I1124 07:12:29.023280 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"655e6590-673e-4fde-a097-1173dfcef9e2","Type":"ContainerDied","Data":"2755fbe8785a1bb670bc1c46dc98b73feee7d48fd6630181c7e1097962fe8d83"} Nov 24 07:12:29 crc kubenswrapper[4809]: I1124 07:12:29.023293 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"655e6590-673e-4fde-a097-1173dfcef9e2","Type":"ContainerDied","Data":"42f9be194d69131769e788b4521d542fb8b90d88ac5fe9d2fc53be376177c837"} Nov 24 07:12:29 crc kubenswrapper[4809]: I1124 07:12:29.062617 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7947c766c8-w6s92"] Nov 24 07:12:29 crc kubenswrapper[4809]: I1124 07:12:29.135902 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 07:12:29 crc kubenswrapper[4809]: I1124 07:12:29.184007 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-xrnts"] Nov 24 07:12:29 crc kubenswrapper[4809]: W1124 07:12:29.190294 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6252aaf6_6d55_47fc_95fb_8b44d9fd652a.slice/crio-e84e1cf61702ac475319ce8d3ac4a4783fd378c445d69f17a4b791b570d5bd08 WatchSource:0}: Error finding container e84e1cf61702ac475319ce8d3ac4a4783fd378c445d69f17a4b791b570d5bd08: Status 404 returned error can't find the container with id e84e1cf61702ac475319ce8d3ac4a4783fd378c445d69f17a4b791b570d5bd08 Nov 24 07:12:29 crc kubenswrapper[4809]: I1124 07:12:29.321104 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-cc9bfd6b9-27jr8"] Nov 24 07:12:29 crc kubenswrapper[4809]: W1124 07:12:29.340174 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c05e3fc_9407_4bc4_86b4_409ebee7f3e7.slice/crio-dd9c8bbe4c95a505a814dcfdd8f99679c094a38859d6bb29a20b7a54c43e223e WatchSource:0}: Error finding container dd9c8bbe4c95a505a814dcfdd8f99679c094a38859d6bb29a20b7a54c43e223e: Status 404 returned error can't find the container with id dd9c8bbe4c95a505a814dcfdd8f99679c094a38859d6bb29a20b7a54c43e223e Nov 24 07:12:29 crc kubenswrapper[4809]: I1124 07:12:29.379844 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6bd6d9b5b4-rvvrx"] Nov 24 07:12:29 crc kubenswrapper[4809]: I1124 07:12:29.446889 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 07:12:29 crc kubenswrapper[4809]: I1124 07:12:29.455190 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-br547"] Nov 24 07:12:29 crc kubenswrapper[4809]: W1124 07:12:29.466507 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6315ca3f_8d9f_4e7e_b326_725025544aed.slice/crio-cac668d4f738fc33d538c49b44eb9fe079c61015eeb6eefa23ec94076a3b5676 WatchSource:0}: Error finding container cac668d4f738fc33d538c49b44eb9fe079c61015eeb6eefa23ec94076a3b5676: Status 404 returned error can't find the container with id cac668d4f738fc33d538c49b44eb9fe079c61015eeb6eefa23ec94076a3b5676 Nov 24 07:12:29 crc kubenswrapper[4809]: I1124 07:12:29.607769 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.044520 4809 generic.go:334] "Generic (PLEG): container finished" podID="ec7ac87d-2d70-4384-aceb-6f2eca6bf94b" containerID="110725378d5569ef98bf845f63c45738e84831d095aa4ac8049621bdef38cd52" exitCode=0 Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.044584 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-br547" event={"ID":"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b","Type":"ContainerDied","Data":"110725378d5569ef98bf845f63c45738e84831d095aa4ac8049621bdef38cd52"} Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.044805 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-br547" event={"ID":"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b","Type":"ContainerStarted","Data":"3ed4b8b95947ba5dffe2a86040a1548d29f83b6a1a0bbda277ef96288140d4d6"} Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.046863 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6315ca3f-8d9f-4e7e-b326-725025544aed","Type":"ContainerStarted","Data":"cac668d4f738fc33d538c49b44eb9fe079c61015eeb6eefa23ec94076a3b5676"} Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.052115 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1a14aa5c-2f38-4db8-9bf9-422c14699895","Type":"ContainerStarted","Data":"ca8cdde3cb833bbe9a883fccc9869306eb85902a1d8464670d26b39ffbea5d08"} Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.063545 4809 generic.go:334] "Generic (PLEG): container finished" podID="6252aaf6-6d55-47fc-95fb-8b44d9fd652a" containerID="a1807aad4be7e6ae14b917e171a9fa1c4243094fa382a0b723dac400f1e4b620" exitCode=0 Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.063619 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-xrnts" event={"ID":"6252aaf6-6d55-47fc-95fb-8b44d9fd652a","Type":"ContainerDied","Data":"a1807aad4be7e6ae14b917e171a9fa1c4243094fa382a0b723dac400f1e4b620"} Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.063650 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-xrnts" event={"ID":"6252aaf6-6d55-47fc-95fb-8b44d9fd652a","Type":"ContainerStarted","Data":"e84e1cf61702ac475319ce8d3ac4a4783fd378c445d69f17a4b791b570d5bd08"} Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.081160 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" event={"ID":"d96aea6e-a61c-437f-b4f5-7cc4db4a6316","Type":"ContainerStarted","Data":"692758c2c3db55c2d181e8b6c7a794d68add17b5b2ebdd723b9bd2c41e55b5ba"} Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.081499 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" event={"ID":"d96aea6e-a61c-437f-b4f5-7cc4db4a6316","Type":"ContainerStarted","Data":"945a8233ed9ab93d359c28ff35a6522c4cac17a32b44e77dbe77c85450d9f51f"} Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.081512 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" event={"ID":"d96aea6e-a61c-437f-b4f5-7cc4db4a6316","Type":"ContainerStarted","Data":"de14891d92e5219342c563ec3fcb62e45330f9b46d37f37ecba024fd0254b5a3"} Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.082258 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.082283 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.096002 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-cc9bfd6b9-27jr8" event={"ID":"5c05e3fc-9407-4bc4-86b4-409ebee7f3e7","Type":"ContainerStarted","Data":"dd9c8bbe4c95a505a814dcfdd8f99679c094a38859d6bb29a20b7a54c43e223e"} Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.099341 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7947c766c8-w6s92" event={"ID":"ada052fc-743e-477f-b5be-848ca3536f09","Type":"ContainerStarted","Data":"4118f4fc8c8c5a9d2489413410c985bb4a5b1cd093d0277f0646a83699c67eb2"} Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.115812 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" podStartSLOduration=2.115792603 podStartE2EDuration="2.115792603s" podCreationTimestamp="2025-11-24 07:12:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:12:30.113187204 +0000 UTC m=+1050.013778909" watchObservedRunningTime="2025-11-24 07:12:30.115792603 +0000 UTC m=+1050.016384298" Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.736023 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.764514 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-xrnts" Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.929353 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-ovsdbserver-sb\") pod \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.929443 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-ovsdbserver-nb\") pod \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.929480 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-config\") pod \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.929602 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-dns-svc\") pod \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.929630 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmf5f\" (UniqueName: \"kubernetes.io/projected/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-kube-api-access-pmf5f\") pod \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.929680 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-dns-swift-storage-0\") pod \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\" (UID: \"6252aaf6-6d55-47fc-95fb-8b44d9fd652a\") " Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.952261 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-kube-api-access-pmf5f" (OuterVolumeSpecName: "kube-api-access-pmf5f") pod "6252aaf6-6d55-47fc-95fb-8b44d9fd652a" (UID: "6252aaf6-6d55-47fc-95fb-8b44d9fd652a"). InnerVolumeSpecName "kube-api-access-pmf5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:12:30 crc kubenswrapper[4809]: I1124 07:12:30.988331 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6252aaf6-6d55-47fc-95fb-8b44d9fd652a" (UID: "6252aaf6-6d55-47fc-95fb-8b44d9fd652a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.014823 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6252aaf6-6d55-47fc-95fb-8b44d9fd652a" (UID: "6252aaf6-6d55-47fc-95fb-8b44d9fd652a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.015416 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6252aaf6-6d55-47fc-95fb-8b44d9fd652a" (UID: "6252aaf6-6d55-47fc-95fb-8b44d9fd652a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.032152 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmf5f\" (UniqueName: \"kubernetes.io/projected/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-kube-api-access-pmf5f\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.032179 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.032188 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.032196 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.073702 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-config" (OuterVolumeSpecName: "config") pod "6252aaf6-6d55-47fc-95fb-8b44d9fd652a" (UID: "6252aaf6-6d55-47fc-95fb-8b44d9fd652a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.123620 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6252aaf6-6d55-47fc-95fb-8b44d9fd652a" (UID: "6252aaf6-6d55-47fc-95fb-8b44d9fd652a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.135158 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.135189 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6252aaf6-6d55-47fc-95fb-8b44d9fd652a-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.141349 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-xrnts" event={"ID":"6252aaf6-6d55-47fc-95fb-8b44d9fd652a","Type":"ContainerDied","Data":"e84e1cf61702ac475319ce8d3ac4a4783fd378c445d69f17a4b791b570d5bd08"} Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.141396 4809 scope.go:117] "RemoveContainer" containerID="a1807aad4be7e6ae14b917e171a9fa1c4243094fa382a0b723dac400f1e4b620" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.141501 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-xrnts" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.154082 4809 generic.go:334] "Generic (PLEG): container finished" podID="655e6590-673e-4fde-a097-1173dfcef9e2" containerID="ffc5ddd380aa10bc65340c8e0facbe2c23a4aba66abadd3808bb97c90888fe85" exitCode=0 Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.154132 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"655e6590-673e-4fde-a097-1173dfcef9e2","Type":"ContainerDied","Data":"ffc5ddd380aa10bc65340c8e0facbe2c23a4aba66abadd3808bb97c90888fe85"} Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.159524 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1a14aa5c-2f38-4db8-9bf9-422c14699895","Type":"ContainerStarted","Data":"fc7217ce9dfe93fc6257d290ff3397837947db3a1b27a521147843c08d49e3e1"} Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.228005 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-xrnts"] Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.235997 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-xrnts"] Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.426112 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.544791 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-combined-ca-bundle\") pod \"655e6590-673e-4fde-a097-1173dfcef9e2\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.544859 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-scripts\") pod \"655e6590-673e-4fde-a097-1173dfcef9e2\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.544895 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhprc\" (UniqueName: \"kubernetes.io/projected/655e6590-673e-4fde-a097-1173dfcef9e2-kube-api-access-mhprc\") pod \"655e6590-673e-4fde-a097-1173dfcef9e2\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.544952 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-sg-core-conf-yaml\") pod \"655e6590-673e-4fde-a097-1173dfcef9e2\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.545103 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/655e6590-673e-4fde-a097-1173dfcef9e2-log-httpd\") pod \"655e6590-673e-4fde-a097-1173dfcef9e2\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.545154 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-config-data\") pod \"655e6590-673e-4fde-a097-1173dfcef9e2\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.545198 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/655e6590-673e-4fde-a097-1173dfcef9e2-run-httpd\") pod \"655e6590-673e-4fde-a097-1173dfcef9e2\" (UID: \"655e6590-673e-4fde-a097-1173dfcef9e2\") " Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.546492 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/655e6590-673e-4fde-a097-1173dfcef9e2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "655e6590-673e-4fde-a097-1173dfcef9e2" (UID: "655e6590-673e-4fde-a097-1173dfcef9e2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.546584 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/655e6590-673e-4fde-a097-1173dfcef9e2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "655e6590-673e-4fde-a097-1173dfcef9e2" (UID: "655e6590-673e-4fde-a097-1173dfcef9e2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.553711 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/655e6590-673e-4fde-a097-1173dfcef9e2-kube-api-access-mhprc" (OuterVolumeSpecName: "kube-api-access-mhprc") pod "655e6590-673e-4fde-a097-1173dfcef9e2" (UID: "655e6590-673e-4fde-a097-1173dfcef9e2"). InnerVolumeSpecName "kube-api-access-mhprc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.554124 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-scripts" (OuterVolumeSpecName: "scripts") pod "655e6590-673e-4fde-a097-1173dfcef9e2" (UID: "655e6590-673e-4fde-a097-1173dfcef9e2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.581083 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "655e6590-673e-4fde-a097-1173dfcef9e2" (UID: "655e6590-673e-4fde-a097-1173dfcef9e2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.647423 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/655e6590-673e-4fde-a097-1173dfcef9e2-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.647784 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.647796 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhprc\" (UniqueName: \"kubernetes.io/projected/655e6590-673e-4fde-a097-1173dfcef9e2-kube-api-access-mhprc\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.647913 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.647940 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/655e6590-673e-4fde-a097-1173dfcef9e2-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.658268 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "655e6590-673e-4fde-a097-1173dfcef9e2" (UID: "655e6590-673e-4fde-a097-1173dfcef9e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.717440 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-config-data" (OuterVolumeSpecName: "config-data") pod "655e6590-673e-4fde-a097-1173dfcef9e2" (UID: "655e6590-673e-4fde-a097-1173dfcef9e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.749409 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:31 crc kubenswrapper[4809]: I1124 07:12:31.749476 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/655e6590-673e-4fde-a097-1173dfcef9e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.178407 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"655e6590-673e-4fde-a097-1173dfcef9e2","Type":"ContainerDied","Data":"b80b1b435358522aa4a781ad61588386f6857b29c7ea99d9f87106d3726e9d0e"} Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.178448 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.178742 4809 scope.go:117] "RemoveContainer" containerID="a64a2f12bea992c607246a7a39249b1ccb7b9647115c0abd676355a3ddaebdb3" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.183731 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-cc9bfd6b9-27jr8" event={"ID":"5c05e3fc-9407-4bc4-86b4-409ebee7f3e7","Type":"ContainerStarted","Data":"a825904dd5f68ef28e5a06f187d88656485e8514f77f3f27737cddd3a769782e"} Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.185486 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7947c766c8-w6s92" event={"ID":"ada052fc-743e-477f-b5be-848ca3536f09","Type":"ContainerStarted","Data":"57f609f73ee617f23a9370170131e79a279a29ec0cea2c5917ce15b2d0aea207"} Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.189983 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-br547" event={"ID":"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b","Type":"ContainerStarted","Data":"a1a35d39a40c6f8fb1fc684916016ba71fa8ba999ca9d9e1802d4c3c9343b4c3"} Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.191037 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.196135 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1a14aa5c-2f38-4db8-9bf9-422c14699895","Type":"ContainerStarted","Data":"4891063f6682c4b2811a0c696b9c08b9dafd90c63f30e8767f8a00f915cc236a"} Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.196348 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.196421 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1a14aa5c-2f38-4db8-9bf9-422c14699895" containerName="cinder-api" containerID="cri-o://4891063f6682c4b2811a0c696b9c08b9dafd90c63f30e8767f8a00f915cc236a" gracePeriod=30 Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.196693 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1a14aa5c-2f38-4db8-9bf9-422c14699895" containerName="cinder-api-log" containerID="cri-o://fc7217ce9dfe93fc6257d290ff3397837947db3a1b27a521147843c08d49e3e1" gracePeriod=30 Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.207368 4809 scope.go:117] "RemoveContainer" containerID="2755fbe8785a1bb670bc1c46dc98b73feee7d48fd6630181c7e1097962fe8d83" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.219418 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-br547" podStartSLOduration=4.219403483 podStartE2EDuration="4.219403483s" podCreationTimestamp="2025-11-24 07:12:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:12:32.214104783 +0000 UTC m=+1052.114696508" watchObservedRunningTime="2025-11-24 07:12:32.219403483 +0000 UTC m=+1052.119995188" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.248696 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.260260 4809 scope.go:117] "RemoveContainer" containerID="ffc5ddd380aa10bc65340c8e0facbe2c23a4aba66abadd3808bb97c90888fe85" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.261572 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.276352 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:12:32 crc kubenswrapper[4809]: E1124 07:12:32.276826 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6252aaf6-6d55-47fc-95fb-8b44d9fd652a" containerName="init" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.276847 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6252aaf6-6d55-47fc-95fb-8b44d9fd652a" containerName="init" Nov 24 07:12:32 crc kubenswrapper[4809]: E1124 07:12:32.276864 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="655e6590-673e-4fde-a097-1173dfcef9e2" containerName="sg-core" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.276874 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="655e6590-673e-4fde-a097-1173dfcef9e2" containerName="sg-core" Nov 24 07:12:32 crc kubenswrapper[4809]: E1124 07:12:32.276888 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="655e6590-673e-4fde-a097-1173dfcef9e2" containerName="proxy-httpd" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.276896 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="655e6590-673e-4fde-a097-1173dfcef9e2" containerName="proxy-httpd" Nov 24 07:12:32 crc kubenswrapper[4809]: E1124 07:12:32.276922 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="655e6590-673e-4fde-a097-1173dfcef9e2" containerName="ceilometer-notification-agent" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.276930 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="655e6590-673e-4fde-a097-1173dfcef9e2" containerName="ceilometer-notification-agent" Nov 24 07:12:32 crc kubenswrapper[4809]: E1124 07:12:32.276952 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="655e6590-673e-4fde-a097-1173dfcef9e2" containerName="ceilometer-central-agent" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.276960 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="655e6590-673e-4fde-a097-1173dfcef9e2" containerName="ceilometer-central-agent" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.277224 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6252aaf6-6d55-47fc-95fb-8b44d9fd652a" containerName="init" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.277244 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="655e6590-673e-4fde-a097-1173dfcef9e2" containerName="ceilometer-notification-agent" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.277262 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="655e6590-673e-4fde-a097-1173dfcef9e2" containerName="proxy-httpd" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.277316 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="655e6590-673e-4fde-a097-1173dfcef9e2" containerName="ceilometer-central-agent" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.277333 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="655e6590-673e-4fde-a097-1173dfcef9e2" containerName="sg-core" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.279246 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.282279 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.283265 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.283246556 podStartE2EDuration="4.283246556s" podCreationTimestamp="2025-11-24 07:12:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:12:32.260492563 +0000 UTC m=+1052.161084268" watchObservedRunningTime="2025-11-24 07:12:32.283246556 +0000 UTC m=+1052.183838261" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.286342 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.304505 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.319225 4809 scope.go:117] "RemoveContainer" containerID="42f9be194d69131769e788b4521d542fb8b90d88ac5fe9d2fc53be376177c837" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.363734 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff37f7ca-34df-422b-98fd-4a63a09c4128-log-httpd\") pod \"ceilometer-0\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.364177 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-config-data\") pod \"ceilometer-0\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.364275 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff37f7ca-34df-422b-98fd-4a63a09c4128-run-httpd\") pod \"ceilometer-0\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.364343 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nf2wm\" (UniqueName: \"kubernetes.io/projected/ff37f7ca-34df-422b-98fd-4a63a09c4128-kube-api-access-nf2wm\") pod \"ceilometer-0\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.364482 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.364559 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.364629 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-scripts\") pod \"ceilometer-0\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.466300 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.466519 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.466587 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-scripts\") pod \"ceilometer-0\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.466731 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff37f7ca-34df-422b-98fd-4a63a09c4128-log-httpd\") pod \"ceilometer-0\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.466865 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-config-data\") pod \"ceilometer-0\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.467394 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff37f7ca-34df-422b-98fd-4a63a09c4128-run-httpd\") pod \"ceilometer-0\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.467476 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nf2wm\" (UniqueName: \"kubernetes.io/projected/ff37f7ca-34df-422b-98fd-4a63a09c4128-kube-api-access-nf2wm\") pod \"ceilometer-0\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.467836 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff37f7ca-34df-422b-98fd-4a63a09c4128-run-httpd\") pod \"ceilometer-0\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.467339 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff37f7ca-34df-422b-98fd-4a63a09c4128-log-httpd\") pod \"ceilometer-0\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.472092 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.473428 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.476630 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-scripts\") pod \"ceilometer-0\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.478652 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-config-data\") pod \"ceilometer-0\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.486553 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nf2wm\" (UniqueName: \"kubernetes.io/projected/ff37f7ca-34df-422b-98fd-4a63a09c4128-kube-api-access-nf2wm\") pod \"ceilometer-0\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.595262 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.947698 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6252aaf6-6d55-47fc-95fb-8b44d9fd652a" path="/var/lib/kubelet/pods/6252aaf6-6d55-47fc-95fb-8b44d9fd652a/volumes" Nov 24 07:12:32 crc kubenswrapper[4809]: I1124 07:12:32.954542 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="655e6590-673e-4fde-a097-1173dfcef9e2" path="/var/lib/kubelet/pods/655e6590-673e-4fde-a097-1173dfcef9e2/volumes" Nov 24 07:12:33 crc kubenswrapper[4809]: I1124 07:12:33.111364 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:12:33 crc kubenswrapper[4809]: I1124 07:12:33.207770 4809 generic.go:334] "Generic (PLEG): container finished" podID="1a14aa5c-2f38-4db8-9bf9-422c14699895" containerID="fc7217ce9dfe93fc6257d290ff3397837947db3a1b27a521147843c08d49e3e1" exitCode=143 Nov 24 07:12:33 crc kubenswrapper[4809]: I1124 07:12:33.207975 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1a14aa5c-2f38-4db8-9bf9-422c14699895","Type":"ContainerDied","Data":"fc7217ce9dfe93fc6257d290ff3397837947db3a1b27a521147843c08d49e3e1"} Nov 24 07:12:33 crc kubenswrapper[4809]: I1124 07:12:33.212993 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff37f7ca-34df-422b-98fd-4a63a09c4128","Type":"ContainerStarted","Data":"7d06d7b89ff91a4b2108d343a8108c87973718ed7ad3de66bfc5088ee9246fad"} Nov 24 07:12:33 crc kubenswrapper[4809]: I1124 07:12:33.214576 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-cc9bfd6b9-27jr8" event={"ID":"5c05e3fc-9407-4bc4-86b4-409ebee7f3e7","Type":"ContainerStarted","Data":"0f16be1420ea2cdae81887e177abf5522d205b1f0816437500f8fa193a5dd928"} Nov 24 07:12:33 crc kubenswrapper[4809]: I1124 07:12:33.219291 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7947c766c8-w6s92" event={"ID":"ada052fc-743e-477f-b5be-848ca3536f09","Type":"ContainerStarted","Data":"36ddcf0a1adbcd7c3d01616d88e8c56aa021b27b80dcea26046fb328c1f102cb"} Nov 24 07:12:33 crc kubenswrapper[4809]: I1124 07:12:33.223119 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6315ca3f-8d9f-4e7e-b326-725025544aed","Type":"ContainerStarted","Data":"07b6ecf43a05f2d37701dab4a7966dcedec7a2b33b1d1aa0c2f9dd9261aabb29"} Nov 24 07:12:33 crc kubenswrapper[4809]: I1124 07:12:33.223148 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6315ca3f-8d9f-4e7e-b326-725025544aed","Type":"ContainerStarted","Data":"8c7c2fc7bfe2501b3f31c9d6c86ad11a668e2430a0c59d131b3cf67d0d9d94c5"} Nov 24 07:12:33 crc kubenswrapper[4809]: I1124 07:12:33.238089 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-cc9bfd6b9-27jr8" podStartSLOduration=4.065942782 podStartE2EDuration="6.23806882s" podCreationTimestamp="2025-11-24 07:12:27 +0000 UTC" firstStartedPulling="2025-11-24 07:12:29.343100187 +0000 UTC m=+1049.243691892" lastFinishedPulling="2025-11-24 07:12:31.515226225 +0000 UTC m=+1051.415817930" observedRunningTime="2025-11-24 07:12:33.231830615 +0000 UTC m=+1053.132422330" watchObservedRunningTime="2025-11-24 07:12:33.23806882 +0000 UTC m=+1053.138660535" Nov 24 07:12:33 crc kubenswrapper[4809]: I1124 07:12:33.256366 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.243725826 podStartE2EDuration="5.256317114s" podCreationTimestamp="2025-11-24 07:12:28 +0000 UTC" firstStartedPulling="2025-11-24 07:12:29.468569904 +0000 UTC m=+1049.369161609" lastFinishedPulling="2025-11-24 07:12:31.481161192 +0000 UTC m=+1051.381752897" observedRunningTime="2025-11-24 07:12:33.254076924 +0000 UTC m=+1053.154668669" watchObservedRunningTime="2025-11-24 07:12:33.256317114 +0000 UTC m=+1053.156908829" Nov 24 07:12:33 crc kubenswrapper[4809]: I1124 07:12:33.301318 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7947c766c8-w6s92" podStartSLOduration=3.881976355 podStartE2EDuration="6.301295886s" podCreationTimestamp="2025-11-24 07:12:27 +0000 UTC" firstStartedPulling="2025-11-24 07:12:29.067568343 +0000 UTC m=+1048.968160048" lastFinishedPulling="2025-11-24 07:12:31.486887874 +0000 UTC m=+1051.387479579" observedRunningTime="2025-11-24 07:12:33.26709475 +0000 UTC m=+1053.167686455" watchObservedRunningTime="2025-11-24 07:12:33.301295886 +0000 UTC m=+1053.201887591" Nov 24 07:12:33 crc kubenswrapper[4809]: I1124 07:12:33.753086 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.237009 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff37f7ca-34df-422b-98fd-4a63a09c4128","Type":"ContainerStarted","Data":"3f927e26283f21a361b4680339fb1242ad429849deb9c5a497e2a9f03e3e28d0"} Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.622480 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7bb85696db-hj9dc"] Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.624277 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.628269 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.628295 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.636860 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7bb85696db-hj9dc"] Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.716274 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04780f37-736e-4f56-a887-e50c11e3b3ac-logs\") pod \"barbican-api-7bb85696db-hj9dc\" (UID: \"04780f37-736e-4f56-a887-e50c11e3b3ac\") " pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.716327 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04780f37-736e-4f56-a887-e50c11e3b3ac-config-data-custom\") pod \"barbican-api-7bb85696db-hj9dc\" (UID: \"04780f37-736e-4f56-a887-e50c11e3b3ac\") " pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.716355 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/04780f37-736e-4f56-a887-e50c11e3b3ac-internal-tls-certs\") pod \"barbican-api-7bb85696db-hj9dc\" (UID: \"04780f37-736e-4f56-a887-e50c11e3b3ac\") " pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.716484 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87vz2\" (UniqueName: \"kubernetes.io/projected/04780f37-736e-4f56-a887-e50c11e3b3ac-kube-api-access-87vz2\") pod \"barbican-api-7bb85696db-hj9dc\" (UID: \"04780f37-736e-4f56-a887-e50c11e3b3ac\") " pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.716560 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04780f37-736e-4f56-a887-e50c11e3b3ac-combined-ca-bundle\") pod \"barbican-api-7bb85696db-hj9dc\" (UID: \"04780f37-736e-4f56-a887-e50c11e3b3ac\") " pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.716833 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04780f37-736e-4f56-a887-e50c11e3b3ac-config-data\") pod \"barbican-api-7bb85696db-hj9dc\" (UID: \"04780f37-736e-4f56-a887-e50c11e3b3ac\") " pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.716891 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/04780f37-736e-4f56-a887-e50c11e3b3ac-public-tls-certs\") pod \"barbican-api-7bb85696db-hj9dc\" (UID: \"04780f37-736e-4f56-a887-e50c11e3b3ac\") " pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.818614 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04780f37-736e-4f56-a887-e50c11e3b3ac-config-data\") pod \"barbican-api-7bb85696db-hj9dc\" (UID: \"04780f37-736e-4f56-a887-e50c11e3b3ac\") " pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.818665 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/04780f37-736e-4f56-a887-e50c11e3b3ac-public-tls-certs\") pod \"barbican-api-7bb85696db-hj9dc\" (UID: \"04780f37-736e-4f56-a887-e50c11e3b3ac\") " pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.818709 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04780f37-736e-4f56-a887-e50c11e3b3ac-logs\") pod \"barbican-api-7bb85696db-hj9dc\" (UID: \"04780f37-736e-4f56-a887-e50c11e3b3ac\") " pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.818739 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04780f37-736e-4f56-a887-e50c11e3b3ac-config-data-custom\") pod \"barbican-api-7bb85696db-hj9dc\" (UID: \"04780f37-736e-4f56-a887-e50c11e3b3ac\") " pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.818757 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/04780f37-736e-4f56-a887-e50c11e3b3ac-internal-tls-certs\") pod \"barbican-api-7bb85696db-hj9dc\" (UID: \"04780f37-736e-4f56-a887-e50c11e3b3ac\") " pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.818783 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87vz2\" (UniqueName: \"kubernetes.io/projected/04780f37-736e-4f56-a887-e50c11e3b3ac-kube-api-access-87vz2\") pod \"barbican-api-7bb85696db-hj9dc\" (UID: \"04780f37-736e-4f56-a887-e50c11e3b3ac\") " pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.818807 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04780f37-736e-4f56-a887-e50c11e3b3ac-combined-ca-bundle\") pod \"barbican-api-7bb85696db-hj9dc\" (UID: \"04780f37-736e-4f56-a887-e50c11e3b3ac\") " pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.819178 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04780f37-736e-4f56-a887-e50c11e3b3ac-logs\") pod \"barbican-api-7bb85696db-hj9dc\" (UID: \"04780f37-736e-4f56-a887-e50c11e3b3ac\") " pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.825588 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/04780f37-736e-4f56-a887-e50c11e3b3ac-internal-tls-certs\") pod \"barbican-api-7bb85696db-hj9dc\" (UID: \"04780f37-736e-4f56-a887-e50c11e3b3ac\") " pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.828263 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04780f37-736e-4f56-a887-e50c11e3b3ac-config-data\") pod \"barbican-api-7bb85696db-hj9dc\" (UID: \"04780f37-736e-4f56-a887-e50c11e3b3ac\") " pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.829793 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04780f37-736e-4f56-a887-e50c11e3b3ac-combined-ca-bundle\") pod \"barbican-api-7bb85696db-hj9dc\" (UID: \"04780f37-736e-4f56-a887-e50c11e3b3ac\") " pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.833534 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/04780f37-736e-4f56-a887-e50c11e3b3ac-public-tls-certs\") pod \"barbican-api-7bb85696db-hj9dc\" (UID: \"04780f37-736e-4f56-a887-e50c11e3b3ac\") " pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.836183 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04780f37-736e-4f56-a887-e50c11e3b3ac-config-data-custom\") pod \"barbican-api-7bb85696db-hj9dc\" (UID: \"04780f37-736e-4f56-a887-e50c11e3b3ac\") " pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.837687 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87vz2\" (UniqueName: \"kubernetes.io/projected/04780f37-736e-4f56-a887-e50c11e3b3ac-kube-api-access-87vz2\") pod \"barbican-api-7bb85696db-hj9dc\" (UID: \"04780f37-736e-4f56-a887-e50c11e3b3ac\") " pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:34 crc kubenswrapper[4809]: I1124 07:12:34.941195 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:35 crc kubenswrapper[4809]: I1124 07:12:35.254246 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff37f7ca-34df-422b-98fd-4a63a09c4128","Type":"ContainerStarted","Data":"1d994e5b49682209744007674d2b4631c46a91caf6c3c6bb5dbe9dcb0f1c4ada"} Nov 24 07:12:35 crc kubenswrapper[4809]: I1124 07:12:35.480211 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7bb85696db-hj9dc"] Nov 24 07:12:36 crc kubenswrapper[4809]: I1124 07:12:36.295947 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bb85696db-hj9dc" event={"ID":"04780f37-736e-4f56-a887-e50c11e3b3ac","Type":"ContainerStarted","Data":"69b030f35afa3214f436a412b7680c815c36003aad3a21eb8f1e0471b1f4380c"} Nov 24 07:12:36 crc kubenswrapper[4809]: I1124 07:12:36.296795 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bb85696db-hj9dc" event={"ID":"04780f37-736e-4f56-a887-e50c11e3b3ac","Type":"ContainerStarted","Data":"4fe8d40c5d71e0e911c5a3d42640adc4c7e4f5d315541fd7f10289fc09bb8026"} Nov 24 07:12:37 crc kubenswrapper[4809]: I1124 07:12:37.305314 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff37f7ca-34df-422b-98fd-4a63a09c4128","Type":"ContainerStarted","Data":"e7205a02519e1179ec022ddcdf182a82755a35878a71ddf9a3a03b759691ccba"} Nov 24 07:12:37 crc kubenswrapper[4809]: I1124 07:12:37.307123 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bb85696db-hj9dc" event={"ID":"04780f37-736e-4f56-a887-e50c11e3b3ac","Type":"ContainerStarted","Data":"9a5eca064d04ab7327dd57edaabfebfde3acc859b1d05ae52b85569e05f15667"} Nov 24 07:12:37 crc kubenswrapper[4809]: I1124 07:12:37.307255 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:37 crc kubenswrapper[4809]: I1124 07:12:37.326597 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:12:37 crc kubenswrapper[4809]: I1124 07:12:37.341835 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7bb85696db-hj9dc" podStartSLOduration=3.341821279 podStartE2EDuration="3.341821279s" podCreationTimestamp="2025-11-24 07:12:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:12:37.338475419 +0000 UTC m=+1057.239067124" watchObservedRunningTime="2025-11-24 07:12:37.341821279 +0000 UTC m=+1057.242412984" Nov 24 07:12:37 crc kubenswrapper[4809]: I1124 07:12:37.618419 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:12:37 crc kubenswrapper[4809]: I1124 07:12:37.991619 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-68b795db4d-rr6j7" Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.324453 4809 generic.go:334] "Generic (PLEG): container finished" podID="46ebff1a-add8-420f-99c2-c593d486d2f8" containerID="e44083aa1ad4fa4b45a13e3c4271aa1d170097c126ada26dcc06b313160fc1c8" exitCode=137 Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.324773 4809 generic.go:334] "Generic (PLEG): container finished" podID="46ebff1a-add8-420f-99c2-c593d486d2f8" containerID="31646cbdae7ff94e4b915bd134d0575d662c1ea9527419ba1ba39e2845859955" exitCode=137 Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.324517 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84666554f7-jc9pp" event={"ID":"46ebff1a-add8-420f-99c2-c593d486d2f8","Type":"ContainerDied","Data":"e44083aa1ad4fa4b45a13e3c4271aa1d170097c126ada26dcc06b313160fc1c8"} Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.324833 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84666554f7-jc9pp" event={"ID":"46ebff1a-add8-420f-99c2-c593d486d2f8","Type":"ContainerDied","Data":"31646cbdae7ff94e4b915bd134d0575d662c1ea9527419ba1ba39e2845859955"} Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.337009 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff37f7ca-34df-422b-98fd-4a63a09c4128","Type":"ContainerStarted","Data":"cdeef0898a6f4fc473a3a31b2881a3fa16c3f2bdd8513e81e2d7fcb1dd6c62fa"} Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.337073 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.337086 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.340392 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-84666554f7-jc9pp" Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.369564 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.651608462 podStartE2EDuration="6.369520564s" podCreationTimestamp="2025-11-24 07:12:32 +0000 UTC" firstStartedPulling="2025-11-24 07:12:33.130672202 +0000 UTC m=+1053.031263917" lastFinishedPulling="2025-11-24 07:12:37.848584294 +0000 UTC m=+1057.749176019" observedRunningTime="2025-11-24 07:12:38.353929571 +0000 UTC m=+1058.254521276" watchObservedRunningTime="2025-11-24 07:12:38.369520564 +0000 UTC m=+1058.270112269" Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.408035 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42s9t\" (UniqueName: \"kubernetes.io/projected/46ebff1a-add8-420f-99c2-c593d486d2f8-kube-api-access-42s9t\") pod \"46ebff1a-add8-420f-99c2-c593d486d2f8\" (UID: \"46ebff1a-add8-420f-99c2-c593d486d2f8\") " Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.408083 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/46ebff1a-add8-420f-99c2-c593d486d2f8-horizon-secret-key\") pod \"46ebff1a-add8-420f-99c2-c593d486d2f8\" (UID: \"46ebff1a-add8-420f-99c2-c593d486d2f8\") " Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.408118 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46ebff1a-add8-420f-99c2-c593d486d2f8-logs\") pod \"46ebff1a-add8-420f-99c2-c593d486d2f8\" (UID: \"46ebff1a-add8-420f-99c2-c593d486d2f8\") " Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.408352 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/46ebff1a-add8-420f-99c2-c593d486d2f8-config-data\") pod \"46ebff1a-add8-420f-99c2-c593d486d2f8\" (UID: \"46ebff1a-add8-420f-99c2-c593d486d2f8\") " Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.408401 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46ebff1a-add8-420f-99c2-c593d486d2f8-scripts\") pod \"46ebff1a-add8-420f-99c2-c593d486d2f8\" (UID: \"46ebff1a-add8-420f-99c2-c593d486d2f8\") " Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.409580 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46ebff1a-add8-420f-99c2-c593d486d2f8-logs" (OuterVolumeSpecName: "logs") pod "46ebff1a-add8-420f-99c2-c593d486d2f8" (UID: "46ebff1a-add8-420f-99c2-c593d486d2f8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.413296 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46ebff1a-add8-420f-99c2-c593d486d2f8-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "46ebff1a-add8-420f-99c2-c593d486d2f8" (UID: "46ebff1a-add8-420f-99c2-c593d486d2f8"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.418442 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46ebff1a-add8-420f-99c2-c593d486d2f8-kube-api-access-42s9t" (OuterVolumeSpecName: "kube-api-access-42s9t") pod "46ebff1a-add8-420f-99c2-c593d486d2f8" (UID: "46ebff1a-add8-420f-99c2-c593d486d2f8"). InnerVolumeSpecName "kube-api-access-42s9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.435382 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46ebff1a-add8-420f-99c2-c593d486d2f8-scripts" (OuterVolumeSpecName: "scripts") pod "46ebff1a-add8-420f-99c2-c593d486d2f8" (UID: "46ebff1a-add8-420f-99c2-c593d486d2f8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.472225 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46ebff1a-add8-420f-99c2-c593d486d2f8-config-data" (OuterVolumeSpecName: "config-data") pod "46ebff1a-add8-420f-99c2-c593d486d2f8" (UID: "46ebff1a-add8-420f-99c2-c593d486d2f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.511225 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42s9t\" (UniqueName: \"kubernetes.io/projected/46ebff1a-add8-420f-99c2-c593d486d2f8-kube-api-access-42s9t\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.511265 4809 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/46ebff1a-add8-420f-99c2-c593d486d2f8-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.511276 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46ebff1a-add8-420f-99c2-c593d486d2f8-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.511286 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/46ebff1a-add8-420f-99c2-c593d486d2f8-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.511295 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46ebff1a-add8-420f-99c2-c593d486d2f8-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.837139 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.932746 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-2pm9c"] Nov 24 07:12:38 crc kubenswrapper[4809]: I1124 07:12:38.933211 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" podUID="07b5d297-b3f9-4b51-a007-045cec29345f" containerName="dnsmasq-dns" containerID="cri-o://d26d77e49721fe7649c0e65807e4597d375dff5e87e3630c335e62e5618c9a30" gracePeriod=10 Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.093395 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-8585547888-qdlwv" Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.132275 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.191048 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.198324 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79cbd445d4-5hkc8"] Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.198572 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-79cbd445d4-5hkc8" podUID="5b3e7ca6-2814-4a14-b1f4-765f8086927e" containerName="horizon-log" containerID="cri-o://1a68297f5bbb4fbec6ca1cf7ed7ec3aaee4d1d4d2c979776b78c71e5790046cc" gracePeriod=30 Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.198965 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-79cbd445d4-5hkc8" podUID="5b3e7ca6-2814-4a14-b1f4-765f8086927e" containerName="horizon" containerID="cri-o://54e0622a23539be0abf29e32ee3cbc72076ffc23105f373a61b7e155e5acb13f" gracePeriod=30 Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.215553 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-79cbd445d4-5hkc8" podUID="5b3e7ca6-2814-4a14-b1f4-765f8086927e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.362336 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-84666554f7-jc9pp" Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.362365 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84666554f7-jc9pp" event={"ID":"46ebff1a-add8-420f-99c2-c593d486d2f8","Type":"ContainerDied","Data":"a8d53e6b56d3b2ddefb6923c5f754fe54a0862ed69f12be9f539be0d5558a4a3"} Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.362426 4809 scope.go:117] "RemoveContainer" containerID="e44083aa1ad4fa4b45a13e3c4271aa1d170097c126ada26dcc06b313160fc1c8" Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.371994 4809 generic.go:334] "Generic (PLEG): container finished" podID="07b5d297-b3f9-4b51-a007-045cec29345f" containerID="d26d77e49721fe7649c0e65807e4597d375dff5e87e3630c335e62e5618c9a30" exitCode=0 Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.372120 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" event={"ID":"07b5d297-b3f9-4b51-a007-045cec29345f","Type":"ContainerDied","Data":"d26d77e49721fe7649c0e65807e4597d375dff5e87e3630c335e62e5618c9a30"} Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.372887 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="6315ca3f-8d9f-4e7e-b326-725025544aed" containerName="cinder-scheduler" containerID="cri-o://8c7c2fc7bfe2501b3f31c9d6c86ad11a668e2430a0c59d131b3cf67d0d9d94c5" gracePeriod=30 Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.373040 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="6315ca3f-8d9f-4e7e-b326-725025544aed" containerName="probe" containerID="cri-o://07b6ecf43a05f2d37701dab4a7966dcedec7a2b33b1d1aa0c2f9dd9261aabb29" gracePeriod=30 Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.407033 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-84666554f7-jc9pp"] Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.417356 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-84666554f7-jc9pp"] Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.622308 4809 scope.go:117] "RemoveContainer" containerID="31646cbdae7ff94e4b915bd134d0575d662c1ea9527419ba1ba39e2845859955" Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.803675 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.948568 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-dns-svc\") pod \"07b5d297-b3f9-4b51-a007-045cec29345f\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.948800 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-ovsdbserver-sb\") pod \"07b5d297-b3f9-4b51-a007-045cec29345f\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.948913 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-ovsdbserver-nb\") pod \"07b5d297-b3f9-4b51-a007-045cec29345f\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.949081 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-dns-swift-storage-0\") pod \"07b5d297-b3f9-4b51-a007-045cec29345f\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.949160 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-config\") pod \"07b5d297-b3f9-4b51-a007-045cec29345f\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.949274 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zls8\" (UniqueName: \"kubernetes.io/projected/07b5d297-b3f9-4b51-a007-045cec29345f-kube-api-access-5zls8\") pod \"07b5d297-b3f9-4b51-a007-045cec29345f\" (UID: \"07b5d297-b3f9-4b51-a007-045cec29345f\") " Nov 24 07:12:39 crc kubenswrapper[4809]: I1124 07:12:39.994504 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07b5d297-b3f9-4b51-a007-045cec29345f-kube-api-access-5zls8" (OuterVolumeSpecName: "kube-api-access-5zls8") pod "07b5d297-b3f9-4b51-a007-045cec29345f" (UID: "07b5d297-b3f9-4b51-a007-045cec29345f"). InnerVolumeSpecName "kube-api-access-5zls8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.054688 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zls8\" (UniqueName: \"kubernetes.io/projected/07b5d297-b3f9-4b51-a007-045cec29345f-kube-api-access-5zls8\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.073876 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "07b5d297-b3f9-4b51-a007-045cec29345f" (UID: "07b5d297-b3f9-4b51-a007-045cec29345f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.078625 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "07b5d297-b3f9-4b51-a007-045cec29345f" (UID: "07b5d297-b3f9-4b51-a007-045cec29345f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.105683 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "07b5d297-b3f9-4b51-a007-045cec29345f" (UID: "07b5d297-b3f9-4b51-a007-045cec29345f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.123335 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-config" (OuterVolumeSpecName: "config") pod "07b5d297-b3f9-4b51-a007-045cec29345f" (UID: "07b5d297-b3f9-4b51-a007-045cec29345f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.134506 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "07b5d297-b3f9-4b51-a007-045cec29345f" (UID: "07b5d297-b3f9-4b51-a007-045cec29345f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.156280 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.156315 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.156324 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.156332 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.156340 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07b5d297-b3f9-4b51-a007-045cec29345f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.384456 4809 generic.go:334] "Generic (PLEG): container finished" podID="6315ca3f-8d9f-4e7e-b326-725025544aed" containerID="07b6ecf43a05f2d37701dab4a7966dcedec7a2b33b1d1aa0c2f9dd9261aabb29" exitCode=0 Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.384526 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6315ca3f-8d9f-4e7e-b326-725025544aed","Type":"ContainerDied","Data":"07b6ecf43a05f2d37701dab4a7966dcedec7a2b33b1d1aa0c2f9dd9261aabb29"} Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.387307 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" event={"ID":"07b5d297-b3f9-4b51-a007-045cec29345f","Type":"ContainerDied","Data":"2e04f5e4b004a048393f233666ce3f0bc76687bb542ed18865168521362cad8a"} Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.387352 4809 scope.go:117] "RemoveContainer" containerID="d26d77e49721fe7649c0e65807e4597d375dff5e87e3630c335e62e5618c9a30" Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.387443 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-2pm9c" Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.405583 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6c5d67c78c-vvvv9" Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.407993 4809 scope.go:117] "RemoveContainer" containerID="a208d8e30cccfa4caedd94d326730c550e78d8541611b8a3823638e6f99ef7be" Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.471028 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-2pm9c"] Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.474816 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.476250 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-68b795db4d-rr6j7"] Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.476417 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-68b795db4d-rr6j7" podUID="923046a7-b198-408c-a853-0c54d8adc84c" containerName="neutron-api" containerID="cri-o://c7abafbc9de95f0d2c23e57deab62dd208867121a2e1920e34a4c6dfde6251b2" gracePeriod=30 Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.476530 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-68b795db4d-rr6j7" podUID="923046a7-b198-408c-a853-0c54d8adc84c" containerName="neutron-httpd" containerID="cri-o://04f0aa5a4dd17cacb5c047a936950427bc739864b143d2d38607b108eb4a6f83" gracePeriod=30 Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.486510 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-2pm9c"] Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.772892 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.901127 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07b5d297-b3f9-4b51-a007-045cec29345f" path="/var/lib/kubelet/pods/07b5d297-b3f9-4b51-a007-045cec29345f/volumes" Nov 24 07:12:40 crc kubenswrapper[4809]: I1124 07:12:40.901806 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46ebff1a-add8-420f-99c2-c593d486d2f8" path="/var/lib/kubelet/pods/46ebff1a-add8-420f-99c2-c593d486d2f8/volumes" Nov 24 07:12:41 crc kubenswrapper[4809]: I1124 07:12:41.406687 4809 generic.go:334] "Generic (PLEG): container finished" podID="923046a7-b198-408c-a853-0c54d8adc84c" containerID="04f0aa5a4dd17cacb5c047a936950427bc739864b143d2d38607b108eb4a6f83" exitCode=0 Nov 24 07:12:41 crc kubenswrapper[4809]: I1124 07:12:41.406677 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68b795db4d-rr6j7" event={"ID":"923046a7-b198-408c-a853-0c54d8adc84c","Type":"ContainerDied","Data":"04f0aa5a4dd17cacb5c047a936950427bc739864b143d2d38607b108eb4a6f83"} Nov 24 07:12:41 crc kubenswrapper[4809]: I1124 07:12:41.604176 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.268037 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.310671 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-combined-ca-bundle\") pod \"6315ca3f-8d9f-4e7e-b326-725025544aed\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.310725 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-scripts\") pod \"6315ca3f-8d9f-4e7e-b326-725025544aed\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.310772 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6315ca3f-8d9f-4e7e-b326-725025544aed-etc-machine-id\") pod \"6315ca3f-8d9f-4e7e-b326-725025544aed\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.310822 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-config-data\") pod \"6315ca3f-8d9f-4e7e-b326-725025544aed\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.310918 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fl4dp\" (UniqueName: \"kubernetes.io/projected/6315ca3f-8d9f-4e7e-b326-725025544aed-kube-api-access-fl4dp\") pod \"6315ca3f-8d9f-4e7e-b326-725025544aed\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.310960 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-config-data-custom\") pod \"6315ca3f-8d9f-4e7e-b326-725025544aed\" (UID: \"6315ca3f-8d9f-4e7e-b326-725025544aed\") " Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.310952 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6315ca3f-8d9f-4e7e-b326-725025544aed-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6315ca3f-8d9f-4e7e-b326-725025544aed" (UID: "6315ca3f-8d9f-4e7e-b326-725025544aed"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.311368 4809 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6315ca3f-8d9f-4e7e-b326-725025544aed-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.319048 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6315ca3f-8d9f-4e7e-b326-725025544aed" (UID: "6315ca3f-8d9f-4e7e-b326-725025544aed"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.319106 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6315ca3f-8d9f-4e7e-b326-725025544aed-kube-api-access-fl4dp" (OuterVolumeSpecName: "kube-api-access-fl4dp") pod "6315ca3f-8d9f-4e7e-b326-725025544aed" (UID: "6315ca3f-8d9f-4e7e-b326-725025544aed"). InnerVolumeSpecName "kube-api-access-fl4dp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.354723 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-scripts" (OuterVolumeSpecName: "scripts") pod "6315ca3f-8d9f-4e7e-b326-725025544aed" (UID: "6315ca3f-8d9f-4e7e-b326-725025544aed"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.391484 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6315ca3f-8d9f-4e7e-b326-725025544aed" (UID: "6315ca3f-8d9f-4e7e-b326-725025544aed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.413205 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.413250 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.413263 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fl4dp\" (UniqueName: \"kubernetes.io/projected/6315ca3f-8d9f-4e7e-b326-725025544aed-kube-api-access-fl4dp\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.413278 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.445658 4809 generic.go:334] "Generic (PLEG): container finished" podID="6315ca3f-8d9f-4e7e-b326-725025544aed" containerID="8c7c2fc7bfe2501b3f31c9d6c86ad11a668e2430a0c59d131b3cf67d0d9d94c5" exitCode=0 Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.445709 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6315ca3f-8d9f-4e7e-b326-725025544aed","Type":"ContainerDied","Data":"8c7c2fc7bfe2501b3f31c9d6c86ad11a668e2430a0c59d131b3cf67d0d9d94c5"} Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.445739 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6315ca3f-8d9f-4e7e-b326-725025544aed","Type":"ContainerDied","Data":"cac668d4f738fc33d538c49b44eb9fe079c61015eeb6eefa23ec94076a3b5676"} Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.445758 4809 scope.go:117] "RemoveContainer" containerID="07b6ecf43a05f2d37701dab4a7966dcedec7a2b33b1d1aa0c2f9dd9261aabb29" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.445932 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.453150 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-config-data" (OuterVolumeSpecName: "config-data") pod "6315ca3f-8d9f-4e7e-b326-725025544aed" (UID: "6315ca3f-8d9f-4e7e-b326-725025544aed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.487150 4809 scope.go:117] "RemoveContainer" containerID="8c7c2fc7bfe2501b3f31c9d6c86ad11a668e2430a0c59d131b3cf67d0d9d94c5" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.514998 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6315ca3f-8d9f-4e7e-b326-725025544aed-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.522635 4809 scope.go:117] "RemoveContainer" containerID="07b6ecf43a05f2d37701dab4a7966dcedec7a2b33b1d1aa0c2f9dd9261aabb29" Nov 24 07:12:43 crc kubenswrapper[4809]: E1124 07:12:43.523146 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07b6ecf43a05f2d37701dab4a7966dcedec7a2b33b1d1aa0c2f9dd9261aabb29\": container with ID starting with 07b6ecf43a05f2d37701dab4a7966dcedec7a2b33b1d1aa0c2f9dd9261aabb29 not found: ID does not exist" containerID="07b6ecf43a05f2d37701dab4a7966dcedec7a2b33b1d1aa0c2f9dd9261aabb29" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.523193 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07b6ecf43a05f2d37701dab4a7966dcedec7a2b33b1d1aa0c2f9dd9261aabb29"} err="failed to get container status \"07b6ecf43a05f2d37701dab4a7966dcedec7a2b33b1d1aa0c2f9dd9261aabb29\": rpc error: code = NotFound desc = could not find container \"07b6ecf43a05f2d37701dab4a7966dcedec7a2b33b1d1aa0c2f9dd9261aabb29\": container with ID starting with 07b6ecf43a05f2d37701dab4a7966dcedec7a2b33b1d1aa0c2f9dd9261aabb29 not found: ID does not exist" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.523223 4809 scope.go:117] "RemoveContainer" containerID="8c7c2fc7bfe2501b3f31c9d6c86ad11a668e2430a0c59d131b3cf67d0d9d94c5" Nov 24 07:12:43 crc kubenswrapper[4809]: E1124 07:12:43.523506 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c7c2fc7bfe2501b3f31c9d6c86ad11a668e2430a0c59d131b3cf67d0d9d94c5\": container with ID starting with 8c7c2fc7bfe2501b3f31c9d6c86ad11a668e2430a0c59d131b3cf67d0d9d94c5 not found: ID does not exist" containerID="8c7c2fc7bfe2501b3f31c9d6c86ad11a668e2430a0c59d131b3cf67d0d9d94c5" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.523534 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c7c2fc7bfe2501b3f31c9d6c86ad11a668e2430a0c59d131b3cf67d0d9d94c5"} err="failed to get container status \"8c7c2fc7bfe2501b3f31c9d6c86ad11a668e2430a0c59d131b3cf67d0d9d94c5\": rpc error: code = NotFound desc = could not find container \"8c7c2fc7bfe2501b3f31c9d6c86ad11a668e2430a0c59d131b3cf67d0d9d94c5\": container with ID starting with 8c7c2fc7bfe2501b3f31c9d6c86ad11a668e2430a0c59d131b3cf67d0d9d94c5 not found: ID does not exist" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.610301 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-79cbd445d4-5hkc8" podUID="5b3e7ca6-2814-4a14-b1f4-765f8086927e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:51626->10.217.0.148:8443: read: connection reset by peer" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.772326 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.782570 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.801746 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 07:12:43 crc kubenswrapper[4809]: E1124 07:12:43.802267 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07b5d297-b3f9-4b51-a007-045cec29345f" containerName="init" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.802284 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="07b5d297-b3f9-4b51-a007-045cec29345f" containerName="init" Nov 24 07:12:43 crc kubenswrapper[4809]: E1124 07:12:43.802297 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6315ca3f-8d9f-4e7e-b326-725025544aed" containerName="cinder-scheduler" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.802303 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6315ca3f-8d9f-4e7e-b326-725025544aed" containerName="cinder-scheduler" Nov 24 07:12:43 crc kubenswrapper[4809]: E1124 07:12:43.802317 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07b5d297-b3f9-4b51-a007-045cec29345f" containerName="dnsmasq-dns" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.802324 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="07b5d297-b3f9-4b51-a007-045cec29345f" containerName="dnsmasq-dns" Nov 24 07:12:43 crc kubenswrapper[4809]: E1124 07:12:43.802343 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46ebff1a-add8-420f-99c2-c593d486d2f8" containerName="horizon" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.802351 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="46ebff1a-add8-420f-99c2-c593d486d2f8" containerName="horizon" Nov 24 07:12:43 crc kubenswrapper[4809]: E1124 07:12:43.802362 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6315ca3f-8d9f-4e7e-b326-725025544aed" containerName="probe" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.802367 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6315ca3f-8d9f-4e7e-b326-725025544aed" containerName="probe" Nov 24 07:12:43 crc kubenswrapper[4809]: E1124 07:12:43.802377 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46ebff1a-add8-420f-99c2-c593d486d2f8" containerName="horizon-log" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.802382 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="46ebff1a-add8-420f-99c2-c593d486d2f8" containerName="horizon-log" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.802550 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="46ebff1a-add8-420f-99c2-c593d486d2f8" containerName="horizon" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.802570 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="07b5d297-b3f9-4b51-a007-045cec29345f" containerName="dnsmasq-dns" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.802581 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6315ca3f-8d9f-4e7e-b326-725025544aed" containerName="probe" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.802589 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6315ca3f-8d9f-4e7e-b326-725025544aed" containerName="cinder-scheduler" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.802602 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="46ebff1a-add8-420f-99c2-c593d486d2f8" containerName="horizon-log" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.803474 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.805948 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.826017 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.919959 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ea4784f-e104-4ad0-873e-7d336cf4d98e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8ea4784f-e104-4ad0-873e-7d336cf4d98e\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.920051 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ea4784f-e104-4ad0-873e-7d336cf4d98e-config-data\") pod \"cinder-scheduler-0\" (UID: \"8ea4784f-e104-4ad0-873e-7d336cf4d98e\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.920076 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ea4784f-e104-4ad0-873e-7d336cf4d98e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8ea4784f-e104-4ad0-873e-7d336cf4d98e\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.920147 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ea4784f-e104-4ad0-873e-7d336cf4d98e-scripts\") pod \"cinder-scheduler-0\" (UID: \"8ea4784f-e104-4ad0-873e-7d336cf4d98e\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.920163 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8ea4784f-e104-4ad0-873e-7d336cf4d98e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8ea4784f-e104-4ad0-873e-7d336cf4d98e\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:43 crc kubenswrapper[4809]: I1124 07:12:43.920198 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlrgx\" (UniqueName: \"kubernetes.io/projected/8ea4784f-e104-4ad0-873e-7d336cf4d98e-kube-api-access-hlrgx\") pod \"cinder-scheduler-0\" (UID: \"8ea4784f-e104-4ad0-873e-7d336cf4d98e\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:44 crc kubenswrapper[4809]: I1124 07:12:44.022823 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ea4784f-e104-4ad0-873e-7d336cf4d98e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8ea4784f-e104-4ad0-873e-7d336cf4d98e\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:44 crc kubenswrapper[4809]: I1124 07:12:44.022944 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ea4784f-e104-4ad0-873e-7d336cf4d98e-config-data\") pod \"cinder-scheduler-0\" (UID: \"8ea4784f-e104-4ad0-873e-7d336cf4d98e\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:44 crc kubenswrapper[4809]: I1124 07:12:44.023010 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ea4784f-e104-4ad0-873e-7d336cf4d98e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8ea4784f-e104-4ad0-873e-7d336cf4d98e\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:44 crc kubenswrapper[4809]: I1124 07:12:44.023050 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ea4784f-e104-4ad0-873e-7d336cf4d98e-scripts\") pod \"cinder-scheduler-0\" (UID: \"8ea4784f-e104-4ad0-873e-7d336cf4d98e\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:44 crc kubenswrapper[4809]: I1124 07:12:44.023068 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8ea4784f-e104-4ad0-873e-7d336cf4d98e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8ea4784f-e104-4ad0-873e-7d336cf4d98e\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:44 crc kubenswrapper[4809]: I1124 07:12:44.023098 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlrgx\" (UniqueName: \"kubernetes.io/projected/8ea4784f-e104-4ad0-873e-7d336cf4d98e-kube-api-access-hlrgx\") pod \"cinder-scheduler-0\" (UID: \"8ea4784f-e104-4ad0-873e-7d336cf4d98e\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:44 crc kubenswrapper[4809]: I1124 07:12:44.024544 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8ea4784f-e104-4ad0-873e-7d336cf4d98e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8ea4784f-e104-4ad0-873e-7d336cf4d98e\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:44 crc kubenswrapper[4809]: I1124 07:12:44.029109 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ea4784f-e104-4ad0-873e-7d336cf4d98e-config-data\") pod \"cinder-scheduler-0\" (UID: \"8ea4784f-e104-4ad0-873e-7d336cf4d98e\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:44 crc kubenswrapper[4809]: I1124 07:12:44.029373 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ea4784f-e104-4ad0-873e-7d336cf4d98e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8ea4784f-e104-4ad0-873e-7d336cf4d98e\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:44 crc kubenswrapper[4809]: I1124 07:12:44.029575 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ea4784f-e104-4ad0-873e-7d336cf4d98e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8ea4784f-e104-4ad0-873e-7d336cf4d98e\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:44 crc kubenswrapper[4809]: I1124 07:12:44.042023 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ea4784f-e104-4ad0-873e-7d336cf4d98e-scripts\") pod \"cinder-scheduler-0\" (UID: \"8ea4784f-e104-4ad0-873e-7d336cf4d98e\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:44 crc kubenswrapper[4809]: I1124 07:12:44.046596 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlrgx\" (UniqueName: \"kubernetes.io/projected/8ea4784f-e104-4ad0-873e-7d336cf4d98e-kube-api-access-hlrgx\") pod \"cinder-scheduler-0\" (UID: \"8ea4784f-e104-4ad0-873e-7d336cf4d98e\") " pod="openstack/cinder-scheduler-0" Nov 24 07:12:44 crc kubenswrapper[4809]: I1124 07:12:44.128682 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 07:12:44 crc kubenswrapper[4809]: I1124 07:12:44.410405 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 07:12:44 crc kubenswrapper[4809]: W1124 07:12:44.416786 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ea4784f_e104_4ad0_873e_7d336cf4d98e.slice/crio-6fe1ec6db3e74dbdc2af9d95868be823d73ac409f53cc5bbc4753c279f8c2630 WatchSource:0}: Error finding container 6fe1ec6db3e74dbdc2af9d95868be823d73ac409f53cc5bbc4753c279f8c2630: Status 404 returned error can't find the container with id 6fe1ec6db3e74dbdc2af9d95868be823d73ac409f53cc5bbc4753c279f8c2630 Nov 24 07:12:44 crc kubenswrapper[4809]: I1124 07:12:44.464882 4809 generic.go:334] "Generic (PLEG): container finished" podID="5b3e7ca6-2814-4a14-b1f4-765f8086927e" containerID="54e0622a23539be0abf29e32ee3cbc72076ffc23105f373a61b7e155e5acb13f" exitCode=0 Nov 24 07:12:44 crc kubenswrapper[4809]: I1124 07:12:44.464947 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79cbd445d4-5hkc8" event={"ID":"5b3e7ca6-2814-4a14-b1f4-765f8086927e","Type":"ContainerDied","Data":"54e0622a23539be0abf29e32ee3cbc72076ffc23105f373a61b7e155e5acb13f"} Nov 24 07:12:44 crc kubenswrapper[4809]: I1124 07:12:44.466187 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8ea4784f-e104-4ad0-873e-7d336cf4d98e","Type":"ContainerStarted","Data":"6fe1ec6db3e74dbdc2af9d95868be823d73ac409f53cc5bbc4753c279f8c2630"} Nov 24 07:12:44 crc kubenswrapper[4809]: I1124 07:12:44.903799 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6315ca3f-8d9f-4e7e-b326-725025544aed" path="/var/lib/kubelet/pods/6315ca3f-8d9f-4e7e-b326-725025544aed/volumes" Nov 24 07:12:45 crc kubenswrapper[4809]: I1124 07:12:45.162317 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-79cbd445d4-5hkc8" podUID="5b3e7ca6-2814-4a14-b1f4-765f8086927e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Nov 24 07:12:45 crc kubenswrapper[4809]: I1124 07:12:45.481652 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8ea4784f-e104-4ad0-873e-7d336cf4d98e","Type":"ContainerStarted","Data":"0adb4ecbbe2694dfb3c231230e94b7e19a8ac52b6eccebc71293589df790fef2"} Nov 24 07:12:46 crc kubenswrapper[4809]: I1124 07:12:46.498179 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8ea4784f-e104-4ad0-873e-7d336cf4d98e","Type":"ContainerStarted","Data":"07bd41af694af542bd5e643a43f85a46fb2825a2b45612c250738864920195fa"} Nov 24 07:12:46 crc kubenswrapper[4809]: I1124 07:12:46.524622 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.524590091 podStartE2EDuration="3.524590091s" podCreationTimestamp="2025-11-24 07:12:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:12:46.520934213 +0000 UTC m=+1066.421525918" watchObservedRunningTime="2025-11-24 07:12:46.524590091 +0000 UTC m=+1066.425181816" Nov 24 07:12:46 crc kubenswrapper[4809]: I1124 07:12:46.729421 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:46 crc kubenswrapper[4809]: I1124 07:12:46.921661 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7bb85696db-hj9dc" Nov 24 07:12:46 crc kubenswrapper[4809]: I1124 07:12:46.998234 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6bd6d9b5b4-rvvrx"] Nov 24 07:12:46 crc kubenswrapper[4809]: I1124 07:12:46.999870 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" podUID="d96aea6e-a61c-437f-b4f5-7cc4db4a6316" containerName="barbican-api-log" containerID="cri-o://945a8233ed9ab93d359c28ff35a6522c4cac17a32b44e77dbe77c85450d9f51f" gracePeriod=30 Nov 24 07:12:47 crc kubenswrapper[4809]: I1124 07:12:47.000057 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" podUID="d96aea6e-a61c-437f-b4f5-7cc4db4a6316" containerName="barbican-api" containerID="cri-o://692758c2c3db55c2d181e8b6c7a794d68add17b5b2ebdd723b9bd2c41e55b5ba" gracePeriod=30 Nov 24 07:12:47 crc kubenswrapper[4809]: I1124 07:12:47.508817 4809 generic.go:334] "Generic (PLEG): container finished" podID="d96aea6e-a61c-437f-b4f5-7cc4db4a6316" containerID="945a8233ed9ab93d359c28ff35a6522c4cac17a32b44e77dbe77c85450d9f51f" exitCode=143 Nov 24 07:12:47 crc kubenswrapper[4809]: I1124 07:12:47.508900 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" event={"ID":"d96aea6e-a61c-437f-b4f5-7cc4db4a6316","Type":"ContainerDied","Data":"945a8233ed9ab93d359c28ff35a6522c4cac17a32b44e77dbe77c85450d9f51f"} Nov 24 07:12:48 crc kubenswrapper[4809]: I1124 07:12:48.043679 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:12:48 crc kubenswrapper[4809]: I1124 07:12:48.044147 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:12:49 crc kubenswrapper[4809]: I1124 07:12:49.129062 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 07:12:49 crc kubenswrapper[4809]: I1124 07:12:49.354242 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 07:12:50 crc kubenswrapper[4809]: I1124 07:12:50.229310 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" podUID="d96aea6e-a61c-437f-b4f5-7cc4db4a6316" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": read tcp 10.217.0.2:44076->10.217.0.161:9311: read: connection reset by peer" Nov 24 07:12:50 crc kubenswrapper[4809]: I1124 07:12:50.229333 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" podUID="d96aea6e-a61c-437f-b4f5-7cc4db4a6316" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": read tcp 10.217.0.2:44066->10.217.0.161:9311: read: connection reset by peer" Nov 24 07:12:50 crc kubenswrapper[4809]: I1124 07:12:50.545816 4809 generic.go:334] "Generic (PLEG): container finished" podID="d96aea6e-a61c-437f-b4f5-7cc4db4a6316" containerID="692758c2c3db55c2d181e8b6c7a794d68add17b5b2ebdd723b9bd2c41e55b5ba" exitCode=0 Nov 24 07:12:50 crc kubenswrapper[4809]: I1124 07:12:50.545858 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" event={"ID":"d96aea6e-a61c-437f-b4f5-7cc4db4a6316","Type":"ContainerDied","Data":"692758c2c3db55c2d181e8b6c7a794d68add17b5b2ebdd723b9bd2c41e55b5ba"} Nov 24 07:12:50 crc kubenswrapper[4809]: I1124 07:12:50.702978 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" Nov 24 07:12:50 crc kubenswrapper[4809]: I1124 07:12:50.866073 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-config-data\") pod \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\" (UID: \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\") " Nov 24 07:12:50 crc kubenswrapper[4809]: I1124 07:12:50.866293 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-combined-ca-bundle\") pod \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\" (UID: \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\") " Nov 24 07:12:50 crc kubenswrapper[4809]: I1124 07:12:50.866343 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-config-data-custom\") pod \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\" (UID: \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\") " Nov 24 07:12:50 crc kubenswrapper[4809]: I1124 07:12:50.866426 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-logs\") pod \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\" (UID: \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\") " Nov 24 07:12:50 crc kubenswrapper[4809]: I1124 07:12:50.866445 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2g58b\" (UniqueName: \"kubernetes.io/projected/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-kube-api-access-2g58b\") pod \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\" (UID: \"d96aea6e-a61c-437f-b4f5-7cc4db4a6316\") " Nov 24 07:12:50 crc kubenswrapper[4809]: I1124 07:12:50.866940 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-logs" (OuterVolumeSpecName: "logs") pod "d96aea6e-a61c-437f-b4f5-7cc4db4a6316" (UID: "d96aea6e-a61c-437f-b4f5-7cc4db4a6316"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:12:50 crc kubenswrapper[4809]: I1124 07:12:50.875132 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-kube-api-access-2g58b" (OuterVolumeSpecName: "kube-api-access-2g58b") pod "d96aea6e-a61c-437f-b4f5-7cc4db4a6316" (UID: "d96aea6e-a61c-437f-b4f5-7cc4db4a6316"). InnerVolumeSpecName "kube-api-access-2g58b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:12:50 crc kubenswrapper[4809]: I1124 07:12:50.875160 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d96aea6e-a61c-437f-b4f5-7cc4db4a6316" (UID: "d96aea6e-a61c-437f-b4f5-7cc4db4a6316"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:50 crc kubenswrapper[4809]: I1124 07:12:50.894941 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d96aea6e-a61c-437f-b4f5-7cc4db4a6316" (UID: "d96aea6e-a61c-437f-b4f5-7cc4db4a6316"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:50 crc kubenswrapper[4809]: I1124 07:12:50.931911 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-config-data" (OuterVolumeSpecName: "config-data") pod "d96aea6e-a61c-437f-b4f5-7cc4db4a6316" (UID: "d96aea6e-a61c-437f-b4f5-7cc4db4a6316"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:50 crc kubenswrapper[4809]: I1124 07:12:50.968847 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:50 crc kubenswrapper[4809]: I1124 07:12:50.968891 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2g58b\" (UniqueName: \"kubernetes.io/projected/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-kube-api-access-2g58b\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:50 crc kubenswrapper[4809]: I1124 07:12:50.968907 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:50 crc kubenswrapper[4809]: I1124 07:12:50.968922 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:50 crc kubenswrapper[4809]: I1124 07:12:50.968937 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d96aea6e-a61c-437f-b4f5-7cc4db4a6316-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:51 crc kubenswrapper[4809]: I1124 07:12:51.363715 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:51 crc kubenswrapper[4809]: I1124 07:12:51.366768 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-78fc658c7d-ntbgd" Nov 24 07:12:51 crc kubenswrapper[4809]: I1124 07:12:51.560161 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" Nov 24 07:12:51 crc kubenswrapper[4809]: I1124 07:12:51.560314 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6bd6d9b5b4-rvvrx" event={"ID":"d96aea6e-a61c-437f-b4f5-7cc4db4a6316","Type":"ContainerDied","Data":"de14891d92e5219342c563ec3fcb62e45330f9b46d37f37ecba024fd0254b5a3"} Nov 24 07:12:51 crc kubenswrapper[4809]: I1124 07:12:51.564394 4809 scope.go:117] "RemoveContainer" containerID="692758c2c3db55c2d181e8b6c7a794d68add17b5b2ebdd723b9bd2c41e55b5ba" Nov 24 07:12:51 crc kubenswrapper[4809]: I1124 07:12:51.603239 4809 scope.go:117] "RemoveContainer" containerID="945a8233ed9ab93d359c28ff35a6522c4cac17a32b44e77dbe77c85450d9f51f" Nov 24 07:12:51 crc kubenswrapper[4809]: I1124 07:12:51.632288 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6bd6d9b5b4-rvvrx"] Nov 24 07:12:51 crc kubenswrapper[4809]: I1124 07:12:51.635603 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-fd8844f9c-6jj95" Nov 24 07:12:51 crc kubenswrapper[4809]: I1124 07:12:51.640590 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6bd6d9b5b4-rvvrx"] Nov 24 07:12:51 crc kubenswrapper[4809]: I1124 07:12:51.824650 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 07:12:51 crc kubenswrapper[4809]: E1124 07:12:51.825055 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d96aea6e-a61c-437f-b4f5-7cc4db4a6316" containerName="barbican-api-log" Nov 24 07:12:51 crc kubenswrapper[4809]: I1124 07:12:51.825070 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d96aea6e-a61c-437f-b4f5-7cc4db4a6316" containerName="barbican-api-log" Nov 24 07:12:51 crc kubenswrapper[4809]: E1124 07:12:51.825079 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d96aea6e-a61c-437f-b4f5-7cc4db4a6316" containerName="barbican-api" Nov 24 07:12:51 crc kubenswrapper[4809]: I1124 07:12:51.825085 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d96aea6e-a61c-437f-b4f5-7cc4db4a6316" containerName="barbican-api" Nov 24 07:12:51 crc kubenswrapper[4809]: I1124 07:12:51.825272 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d96aea6e-a61c-437f-b4f5-7cc4db4a6316" containerName="barbican-api-log" Nov 24 07:12:51 crc kubenswrapper[4809]: I1124 07:12:51.825300 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d96aea6e-a61c-437f-b4f5-7cc4db4a6316" containerName="barbican-api" Nov 24 07:12:51 crc kubenswrapper[4809]: I1124 07:12:51.826140 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 07:12:51 crc kubenswrapper[4809]: I1124 07:12:51.828463 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-5dnk7" Nov 24 07:12:51 crc kubenswrapper[4809]: I1124 07:12:51.828930 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 24 07:12:51 crc kubenswrapper[4809]: I1124 07:12:51.828942 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 24 07:12:51 crc kubenswrapper[4809]: I1124 07:12:51.847626 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 07:12:51 crc kubenswrapper[4809]: I1124 07:12:51.990891 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/45b7caaf-bebc-4bef-804b-a39246b436a0-openstack-config\") pod \"openstackclient\" (UID: \"45b7caaf-bebc-4bef-804b-a39246b436a0\") " pod="openstack/openstackclient" Nov 24 07:12:51 crc kubenswrapper[4809]: I1124 07:12:51.991008 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/45b7caaf-bebc-4bef-804b-a39246b436a0-openstack-config-secret\") pod \"openstackclient\" (UID: \"45b7caaf-bebc-4bef-804b-a39246b436a0\") " pod="openstack/openstackclient" Nov 24 07:12:51 crc kubenswrapper[4809]: I1124 07:12:51.991495 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45b7caaf-bebc-4bef-804b-a39246b436a0-combined-ca-bundle\") pod \"openstackclient\" (UID: \"45b7caaf-bebc-4bef-804b-a39246b436a0\") " pod="openstack/openstackclient" Nov 24 07:12:51 crc kubenswrapper[4809]: I1124 07:12:51.991520 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k99kc\" (UniqueName: \"kubernetes.io/projected/45b7caaf-bebc-4bef-804b-a39246b436a0-kube-api-access-k99kc\") pod \"openstackclient\" (UID: \"45b7caaf-bebc-4bef-804b-a39246b436a0\") " pod="openstack/openstackclient" Nov 24 07:12:52 crc kubenswrapper[4809]: I1124 07:12:52.094494 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/45b7caaf-bebc-4bef-804b-a39246b436a0-openstack-config\") pod \"openstackclient\" (UID: \"45b7caaf-bebc-4bef-804b-a39246b436a0\") " pod="openstack/openstackclient" Nov 24 07:12:52 crc kubenswrapper[4809]: I1124 07:12:52.094570 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/45b7caaf-bebc-4bef-804b-a39246b436a0-openstack-config-secret\") pod \"openstackclient\" (UID: \"45b7caaf-bebc-4bef-804b-a39246b436a0\") " pod="openstack/openstackclient" Nov 24 07:12:52 crc kubenswrapper[4809]: I1124 07:12:52.094618 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45b7caaf-bebc-4bef-804b-a39246b436a0-combined-ca-bundle\") pod \"openstackclient\" (UID: \"45b7caaf-bebc-4bef-804b-a39246b436a0\") " pod="openstack/openstackclient" Nov 24 07:12:52 crc kubenswrapper[4809]: I1124 07:12:52.095444 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k99kc\" (UniqueName: \"kubernetes.io/projected/45b7caaf-bebc-4bef-804b-a39246b436a0-kube-api-access-k99kc\") pod \"openstackclient\" (UID: \"45b7caaf-bebc-4bef-804b-a39246b436a0\") " pod="openstack/openstackclient" Nov 24 07:12:52 crc kubenswrapper[4809]: I1124 07:12:52.095594 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/45b7caaf-bebc-4bef-804b-a39246b436a0-openstack-config\") pod \"openstackclient\" (UID: \"45b7caaf-bebc-4bef-804b-a39246b436a0\") " pod="openstack/openstackclient" Nov 24 07:12:52 crc kubenswrapper[4809]: I1124 07:12:52.100926 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45b7caaf-bebc-4bef-804b-a39246b436a0-combined-ca-bundle\") pod \"openstackclient\" (UID: \"45b7caaf-bebc-4bef-804b-a39246b436a0\") " pod="openstack/openstackclient" Nov 24 07:12:52 crc kubenswrapper[4809]: I1124 07:12:52.110527 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/45b7caaf-bebc-4bef-804b-a39246b436a0-openstack-config-secret\") pod \"openstackclient\" (UID: \"45b7caaf-bebc-4bef-804b-a39246b436a0\") " pod="openstack/openstackclient" Nov 24 07:12:52 crc kubenswrapper[4809]: I1124 07:12:52.111100 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k99kc\" (UniqueName: \"kubernetes.io/projected/45b7caaf-bebc-4bef-804b-a39246b436a0-kube-api-access-k99kc\") pod \"openstackclient\" (UID: \"45b7caaf-bebc-4bef-804b-a39246b436a0\") " pod="openstack/openstackclient" Nov 24 07:12:52 crc kubenswrapper[4809]: I1124 07:12:52.151620 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 07:12:52 crc kubenswrapper[4809]: I1124 07:12:52.628631 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 07:12:52 crc kubenswrapper[4809]: I1124 07:12:52.920272 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d96aea6e-a61c-437f-b4f5-7cc4db4a6316" path="/var/lib/kubelet/pods/d96aea6e-a61c-437f-b4f5-7cc4db4a6316/volumes" Nov 24 07:12:52 crc kubenswrapper[4809]: I1124 07:12:52.996799 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-68b795db4d-rr6j7" Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.112119 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-ovndb-tls-certs\") pod \"923046a7-b198-408c-a853-0c54d8adc84c\" (UID: \"923046a7-b198-408c-a853-0c54d8adc84c\") " Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.112377 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-config\") pod \"923046a7-b198-408c-a853-0c54d8adc84c\" (UID: \"923046a7-b198-408c-a853-0c54d8adc84c\") " Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.112484 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-combined-ca-bundle\") pod \"923046a7-b198-408c-a853-0c54d8adc84c\" (UID: \"923046a7-b198-408c-a853-0c54d8adc84c\") " Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.112605 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhhjr\" (UniqueName: \"kubernetes.io/projected/923046a7-b198-408c-a853-0c54d8adc84c-kube-api-access-hhhjr\") pod \"923046a7-b198-408c-a853-0c54d8adc84c\" (UID: \"923046a7-b198-408c-a853-0c54d8adc84c\") " Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.112706 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-httpd-config\") pod \"923046a7-b198-408c-a853-0c54d8adc84c\" (UID: \"923046a7-b198-408c-a853-0c54d8adc84c\") " Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.119443 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "923046a7-b198-408c-a853-0c54d8adc84c" (UID: "923046a7-b198-408c-a853-0c54d8adc84c"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.119554 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/923046a7-b198-408c-a853-0c54d8adc84c-kube-api-access-hhhjr" (OuterVolumeSpecName: "kube-api-access-hhhjr") pod "923046a7-b198-408c-a853-0c54d8adc84c" (UID: "923046a7-b198-408c-a853-0c54d8adc84c"). InnerVolumeSpecName "kube-api-access-hhhjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.168246 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "923046a7-b198-408c-a853-0c54d8adc84c" (UID: "923046a7-b198-408c-a853-0c54d8adc84c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.181474 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-config" (OuterVolumeSpecName: "config") pod "923046a7-b198-408c-a853-0c54d8adc84c" (UID: "923046a7-b198-408c-a853-0c54d8adc84c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.197956 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "923046a7-b198-408c-a853-0c54d8adc84c" (UID: "923046a7-b198-408c-a853-0c54d8adc84c"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.214489 4809 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.214531 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.214544 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.214555 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhhjr\" (UniqueName: \"kubernetes.io/projected/923046a7-b198-408c-a853-0c54d8adc84c-kube-api-access-hhhjr\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.214570 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/923046a7-b198-408c-a853-0c54d8adc84c-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.584058 4809 generic.go:334] "Generic (PLEG): container finished" podID="923046a7-b198-408c-a853-0c54d8adc84c" containerID="c7abafbc9de95f0d2c23e57deab62dd208867121a2e1920e34a4c6dfde6251b2" exitCode=0 Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.584115 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68b795db4d-rr6j7" event={"ID":"923046a7-b198-408c-a853-0c54d8adc84c","Type":"ContainerDied","Data":"c7abafbc9de95f0d2c23e57deab62dd208867121a2e1920e34a4c6dfde6251b2"} Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.584141 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68b795db4d-rr6j7" event={"ID":"923046a7-b198-408c-a853-0c54d8adc84c","Type":"ContainerDied","Data":"0d0abc01fe8dad3163a0732353cd8ebc107da004248fbfeb459246573f397e6c"} Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.584157 4809 scope.go:117] "RemoveContainer" containerID="04f0aa5a4dd17cacb5c047a936950427bc739864b143d2d38607b108eb4a6f83" Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.584232 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-68b795db4d-rr6j7" Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.594410 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"45b7caaf-bebc-4bef-804b-a39246b436a0","Type":"ContainerStarted","Data":"3882d8558af91b894cf8f29075bd98344e0e6a75657769509c84cf10902ba300"} Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.669423 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-68b795db4d-rr6j7"] Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.669733 4809 scope.go:117] "RemoveContainer" containerID="c7abafbc9de95f0d2c23e57deab62dd208867121a2e1920e34a4c6dfde6251b2" Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.674463 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-68b795db4d-rr6j7"] Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.708039 4809 scope.go:117] "RemoveContainer" containerID="04f0aa5a4dd17cacb5c047a936950427bc739864b143d2d38607b108eb4a6f83" Nov 24 07:12:53 crc kubenswrapper[4809]: E1124 07:12:53.708785 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04f0aa5a4dd17cacb5c047a936950427bc739864b143d2d38607b108eb4a6f83\": container with ID starting with 04f0aa5a4dd17cacb5c047a936950427bc739864b143d2d38607b108eb4a6f83 not found: ID does not exist" containerID="04f0aa5a4dd17cacb5c047a936950427bc739864b143d2d38607b108eb4a6f83" Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.708824 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04f0aa5a4dd17cacb5c047a936950427bc739864b143d2d38607b108eb4a6f83"} err="failed to get container status \"04f0aa5a4dd17cacb5c047a936950427bc739864b143d2d38607b108eb4a6f83\": rpc error: code = NotFound desc = could not find container \"04f0aa5a4dd17cacb5c047a936950427bc739864b143d2d38607b108eb4a6f83\": container with ID starting with 04f0aa5a4dd17cacb5c047a936950427bc739864b143d2d38607b108eb4a6f83 not found: ID does not exist" Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.708852 4809 scope.go:117] "RemoveContainer" containerID="c7abafbc9de95f0d2c23e57deab62dd208867121a2e1920e34a4c6dfde6251b2" Nov 24 07:12:53 crc kubenswrapper[4809]: E1124 07:12:53.709134 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7abafbc9de95f0d2c23e57deab62dd208867121a2e1920e34a4c6dfde6251b2\": container with ID starting with c7abafbc9de95f0d2c23e57deab62dd208867121a2e1920e34a4c6dfde6251b2 not found: ID does not exist" containerID="c7abafbc9de95f0d2c23e57deab62dd208867121a2e1920e34a4c6dfde6251b2" Nov 24 07:12:53 crc kubenswrapper[4809]: I1124 07:12:53.709173 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7abafbc9de95f0d2c23e57deab62dd208867121a2e1920e34a4c6dfde6251b2"} err="failed to get container status \"c7abafbc9de95f0d2c23e57deab62dd208867121a2e1920e34a4c6dfde6251b2\": rpc error: code = NotFound desc = could not find container \"c7abafbc9de95f0d2c23e57deab62dd208867121a2e1920e34a4c6dfde6251b2\": container with ID starting with c7abafbc9de95f0d2c23e57deab62dd208867121a2e1920e34a4c6dfde6251b2 not found: ID does not exist" Nov 24 07:12:54 crc kubenswrapper[4809]: I1124 07:12:54.901893 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="923046a7-b198-408c-a853-0c54d8adc84c" path="/var/lib/kubelet/pods/923046a7-b198-408c-a853-0c54d8adc84c/volumes" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.161628 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-79cbd445d4-5hkc8" podUID="5b3e7ca6-2814-4a14-b1f4-765f8086927e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.649771 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-6d74d65d8c-rgmvx"] Nov 24 07:12:55 crc kubenswrapper[4809]: E1124 07:12:55.650547 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="923046a7-b198-408c-a853-0c54d8adc84c" containerName="neutron-httpd" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.650569 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="923046a7-b198-408c-a853-0c54d8adc84c" containerName="neutron-httpd" Nov 24 07:12:55 crc kubenswrapper[4809]: E1124 07:12:55.650617 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="923046a7-b198-408c-a853-0c54d8adc84c" containerName="neutron-api" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.650625 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="923046a7-b198-408c-a853-0c54d8adc84c" containerName="neutron-api" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.650816 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="923046a7-b198-408c-a853-0c54d8adc84c" containerName="neutron-api" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.650853 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="923046a7-b198-408c-a853-0c54d8adc84c" containerName="neutron-httpd" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.652094 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.659165 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.659465 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.661232 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c1fe2a7-0b4f-46ee-8368-624499f8c095-public-tls-certs\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.661279 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c1fe2a7-0b4f-46ee-8368-624499f8c095-config-data\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.661400 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c1fe2a7-0b4f-46ee-8368-624499f8c095-log-httpd\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.661432 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgdtx\" (UniqueName: \"kubernetes.io/projected/2c1fe2a7-0b4f-46ee-8368-624499f8c095-kube-api-access-sgdtx\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.661455 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2c1fe2a7-0b4f-46ee-8368-624499f8c095-etc-swift\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.661560 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c1fe2a7-0b4f-46ee-8368-624499f8c095-internal-tls-certs\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.661612 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c1fe2a7-0b4f-46ee-8368-624499f8c095-combined-ca-bundle\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.661649 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.661668 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c1fe2a7-0b4f-46ee-8368-624499f8c095-run-httpd\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.674415 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6d74d65d8c-rgmvx"] Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.763347 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c1fe2a7-0b4f-46ee-8368-624499f8c095-public-tls-certs\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.763404 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c1fe2a7-0b4f-46ee-8368-624499f8c095-config-data\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.763459 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c1fe2a7-0b4f-46ee-8368-624499f8c095-log-httpd\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.763480 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgdtx\" (UniqueName: \"kubernetes.io/projected/2c1fe2a7-0b4f-46ee-8368-624499f8c095-kube-api-access-sgdtx\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.763503 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2c1fe2a7-0b4f-46ee-8368-624499f8c095-etc-swift\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.763528 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c1fe2a7-0b4f-46ee-8368-624499f8c095-internal-tls-certs\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.763545 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c1fe2a7-0b4f-46ee-8368-624499f8c095-combined-ca-bundle\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.763566 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c1fe2a7-0b4f-46ee-8368-624499f8c095-run-httpd\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.764075 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c1fe2a7-0b4f-46ee-8368-624499f8c095-run-httpd\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.764348 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c1fe2a7-0b4f-46ee-8368-624499f8c095-log-httpd\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.770912 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c1fe2a7-0b4f-46ee-8368-624499f8c095-internal-tls-certs\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.772817 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c1fe2a7-0b4f-46ee-8368-624499f8c095-public-tls-certs\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.773492 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2c1fe2a7-0b4f-46ee-8368-624499f8c095-etc-swift\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.780726 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c1fe2a7-0b4f-46ee-8368-624499f8c095-config-data\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.785022 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgdtx\" (UniqueName: \"kubernetes.io/projected/2c1fe2a7-0b4f-46ee-8368-624499f8c095-kube-api-access-sgdtx\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.795791 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c1fe2a7-0b4f-46ee-8368-624499f8c095-combined-ca-bundle\") pod \"swift-proxy-6d74d65d8c-rgmvx\" (UID: \"2c1fe2a7-0b4f-46ee-8368-624499f8c095\") " pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:55 crc kubenswrapper[4809]: I1124 07:12:55.975565 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:12:56 crc kubenswrapper[4809]: I1124 07:12:56.502677 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6d74d65d8c-rgmvx"] Nov 24 07:12:56 crc kubenswrapper[4809]: I1124 07:12:56.549417 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:12:56 crc kubenswrapper[4809]: I1124 07:12:56.549697 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ff37f7ca-34df-422b-98fd-4a63a09c4128" containerName="ceilometer-central-agent" containerID="cri-o://3f927e26283f21a361b4680339fb1242ad429849deb9c5a497e2a9f03e3e28d0" gracePeriod=30 Nov 24 07:12:56 crc kubenswrapper[4809]: I1124 07:12:56.549727 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ff37f7ca-34df-422b-98fd-4a63a09c4128" containerName="proxy-httpd" containerID="cri-o://cdeef0898a6f4fc473a3a31b2881a3fa16c3f2bdd8513e81e2d7fcb1dd6c62fa" gracePeriod=30 Nov 24 07:12:56 crc kubenswrapper[4809]: I1124 07:12:56.549749 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ff37f7ca-34df-422b-98fd-4a63a09c4128" containerName="sg-core" containerID="cri-o://e7205a02519e1179ec022ddcdf182a82755a35878a71ddf9a3a03b759691ccba" gracePeriod=30 Nov 24 07:12:56 crc kubenswrapper[4809]: I1124 07:12:56.549807 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ff37f7ca-34df-422b-98fd-4a63a09c4128" containerName="ceilometer-notification-agent" containerID="cri-o://1d994e5b49682209744007674d2b4631c46a91caf6c3c6bb5dbe9dcb0f1c4ada" gracePeriod=30 Nov 24 07:12:56 crc kubenswrapper[4809]: I1124 07:12:56.656635 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="ff37f7ca-34df-422b-98fd-4a63a09c4128" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.165:3000/\": read tcp 10.217.0.2:48394->10.217.0.165:3000: read: connection reset by peer" Nov 24 07:12:56 crc kubenswrapper[4809]: I1124 07:12:56.904707 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-q8dsb"] Nov 24 07:12:56 crc kubenswrapper[4809]: I1124 07:12:56.905711 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-q8dsb" Nov 24 07:12:56 crc kubenswrapper[4809]: I1124 07:12:56.936018 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-q8dsb"] Nov 24 07:12:56 crc kubenswrapper[4809]: I1124 07:12:56.991056 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/603e7b5e-01e3-4016-a1bb-9842c9e9ae1e-operator-scripts\") pod \"nova-api-db-create-q8dsb\" (UID: \"603e7b5e-01e3-4016-a1bb-9842c9e9ae1e\") " pod="openstack/nova-api-db-create-q8dsb" Nov 24 07:12:56 crc kubenswrapper[4809]: I1124 07:12:56.991173 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk64j\" (UniqueName: \"kubernetes.io/projected/603e7b5e-01e3-4016-a1bb-9842c9e9ae1e-kube-api-access-wk64j\") pod \"nova-api-db-create-q8dsb\" (UID: \"603e7b5e-01e3-4016-a1bb-9842c9e9ae1e\") " pod="openstack/nova-api-db-create-q8dsb" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.002226 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-wn4xp"] Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.003691 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wn4xp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.023089 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-wn4xp"] Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.039002 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-6e9f-account-create-7788v"] Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.040218 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6e9f-account-create-7788v" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.044684 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.056599 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-6e9f-account-create-7788v"] Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.093027 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4197d15d-b7aa-4efe-ad80-c3d731afbc44-operator-scripts\") pod \"nova-api-6e9f-account-create-7788v\" (UID: \"4197d15d-b7aa-4efe-ad80-c3d731afbc44\") " pod="openstack/nova-api-6e9f-account-create-7788v" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.093087 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d43a2b0-ce7c-4392-a104-74b2a2a725c1-operator-scripts\") pod \"nova-cell0-db-create-wn4xp\" (UID: \"7d43a2b0-ce7c-4392-a104-74b2a2a725c1\") " pod="openstack/nova-cell0-db-create-wn4xp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.093108 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68q7k\" (UniqueName: \"kubernetes.io/projected/7d43a2b0-ce7c-4392-a104-74b2a2a725c1-kube-api-access-68q7k\") pod \"nova-cell0-db-create-wn4xp\" (UID: \"7d43a2b0-ce7c-4392-a104-74b2a2a725c1\") " pod="openstack/nova-cell0-db-create-wn4xp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.093153 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgttz\" (UniqueName: \"kubernetes.io/projected/4197d15d-b7aa-4efe-ad80-c3d731afbc44-kube-api-access-jgttz\") pod \"nova-api-6e9f-account-create-7788v\" (UID: \"4197d15d-b7aa-4efe-ad80-c3d731afbc44\") " pod="openstack/nova-api-6e9f-account-create-7788v" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.093198 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/603e7b5e-01e3-4016-a1bb-9842c9e9ae1e-operator-scripts\") pod \"nova-api-db-create-q8dsb\" (UID: \"603e7b5e-01e3-4016-a1bb-9842c9e9ae1e\") " pod="openstack/nova-api-db-create-q8dsb" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.093275 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk64j\" (UniqueName: \"kubernetes.io/projected/603e7b5e-01e3-4016-a1bb-9842c9e9ae1e-kube-api-access-wk64j\") pod \"nova-api-db-create-q8dsb\" (UID: \"603e7b5e-01e3-4016-a1bb-9842c9e9ae1e\") " pod="openstack/nova-api-db-create-q8dsb" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.094771 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/603e7b5e-01e3-4016-a1bb-9842c9e9ae1e-operator-scripts\") pod \"nova-api-db-create-q8dsb\" (UID: \"603e7b5e-01e3-4016-a1bb-9842c9e9ae1e\") " pod="openstack/nova-api-db-create-q8dsb" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.115160 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-l7mxp"] Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.116823 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-l7mxp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.124991 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk64j\" (UniqueName: \"kubernetes.io/projected/603e7b5e-01e3-4016-a1bb-9842c9e9ae1e-kube-api-access-wk64j\") pod \"nova-api-db-create-q8dsb\" (UID: \"603e7b5e-01e3-4016-a1bb-9842c9e9ae1e\") " pod="openstack/nova-api-db-create-q8dsb" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.131688 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-l7mxp"] Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.194498 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25b3b27c-bf48-4067-b9fa-2312ff79fe66-operator-scripts\") pod \"nova-cell1-db-create-l7mxp\" (UID: \"25b3b27c-bf48-4067-b9fa-2312ff79fe66\") " pod="openstack/nova-cell1-db-create-l7mxp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.194552 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4197d15d-b7aa-4efe-ad80-c3d731afbc44-operator-scripts\") pod \"nova-api-6e9f-account-create-7788v\" (UID: \"4197d15d-b7aa-4efe-ad80-c3d731afbc44\") " pod="openstack/nova-api-6e9f-account-create-7788v" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.194577 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d43a2b0-ce7c-4392-a104-74b2a2a725c1-operator-scripts\") pod \"nova-cell0-db-create-wn4xp\" (UID: \"7d43a2b0-ce7c-4392-a104-74b2a2a725c1\") " pod="openstack/nova-cell0-db-create-wn4xp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.194593 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68q7k\" (UniqueName: \"kubernetes.io/projected/7d43a2b0-ce7c-4392-a104-74b2a2a725c1-kube-api-access-68q7k\") pod \"nova-cell0-db-create-wn4xp\" (UID: \"7d43a2b0-ce7c-4392-a104-74b2a2a725c1\") " pod="openstack/nova-cell0-db-create-wn4xp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.194611 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvjk2\" (UniqueName: \"kubernetes.io/projected/25b3b27c-bf48-4067-b9fa-2312ff79fe66-kube-api-access-xvjk2\") pod \"nova-cell1-db-create-l7mxp\" (UID: \"25b3b27c-bf48-4067-b9fa-2312ff79fe66\") " pod="openstack/nova-cell1-db-create-l7mxp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.194649 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgttz\" (UniqueName: \"kubernetes.io/projected/4197d15d-b7aa-4efe-ad80-c3d731afbc44-kube-api-access-jgttz\") pod \"nova-api-6e9f-account-create-7788v\" (UID: \"4197d15d-b7aa-4efe-ad80-c3d731afbc44\") " pod="openstack/nova-api-6e9f-account-create-7788v" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.195562 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4197d15d-b7aa-4efe-ad80-c3d731afbc44-operator-scripts\") pod \"nova-api-6e9f-account-create-7788v\" (UID: \"4197d15d-b7aa-4efe-ad80-c3d731afbc44\") " pod="openstack/nova-api-6e9f-account-create-7788v" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.196406 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d43a2b0-ce7c-4392-a104-74b2a2a725c1-operator-scripts\") pod \"nova-cell0-db-create-wn4xp\" (UID: \"7d43a2b0-ce7c-4392-a104-74b2a2a725c1\") " pod="openstack/nova-cell0-db-create-wn4xp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.215662 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68q7k\" (UniqueName: \"kubernetes.io/projected/7d43a2b0-ce7c-4392-a104-74b2a2a725c1-kube-api-access-68q7k\") pod \"nova-cell0-db-create-wn4xp\" (UID: \"7d43a2b0-ce7c-4392-a104-74b2a2a725c1\") " pod="openstack/nova-cell0-db-create-wn4xp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.216068 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-4079-account-create-7bgmp"] Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.217116 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-4079-account-create-7bgmp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.220001 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.228678 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-q8dsb" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.229919 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-4079-account-create-7bgmp"] Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.234675 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgttz\" (UniqueName: \"kubernetes.io/projected/4197d15d-b7aa-4efe-ad80-c3d731afbc44-kube-api-access-jgttz\") pod \"nova-api-6e9f-account-create-7788v\" (UID: \"4197d15d-b7aa-4efe-ad80-c3d731afbc44\") " pod="openstack/nova-api-6e9f-account-create-7788v" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.297737 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25b3b27c-bf48-4067-b9fa-2312ff79fe66-operator-scripts\") pod \"nova-cell1-db-create-l7mxp\" (UID: \"25b3b27c-bf48-4067-b9fa-2312ff79fe66\") " pod="openstack/nova-cell1-db-create-l7mxp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.297790 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvjk2\" (UniqueName: \"kubernetes.io/projected/25b3b27c-bf48-4067-b9fa-2312ff79fe66-kube-api-access-xvjk2\") pod \"nova-cell1-db-create-l7mxp\" (UID: \"25b3b27c-bf48-4067-b9fa-2312ff79fe66\") " pod="openstack/nova-cell1-db-create-l7mxp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.297835 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01ab5f01-bc77-40d0-aacc-3af0e5c09e3f-operator-scripts\") pod \"nova-cell0-4079-account-create-7bgmp\" (UID: \"01ab5f01-bc77-40d0-aacc-3af0e5c09e3f\") " pod="openstack/nova-cell0-4079-account-create-7bgmp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.297874 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm55t\" (UniqueName: \"kubernetes.io/projected/01ab5f01-bc77-40d0-aacc-3af0e5c09e3f-kube-api-access-gm55t\") pod \"nova-cell0-4079-account-create-7bgmp\" (UID: \"01ab5f01-bc77-40d0-aacc-3af0e5c09e3f\") " pod="openstack/nova-cell0-4079-account-create-7bgmp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.298571 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25b3b27c-bf48-4067-b9fa-2312ff79fe66-operator-scripts\") pod \"nova-cell1-db-create-l7mxp\" (UID: \"25b3b27c-bf48-4067-b9fa-2312ff79fe66\") " pod="openstack/nova-cell1-db-create-l7mxp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.315502 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvjk2\" (UniqueName: \"kubernetes.io/projected/25b3b27c-bf48-4067-b9fa-2312ff79fe66-kube-api-access-xvjk2\") pod \"nova-cell1-db-create-l7mxp\" (UID: \"25b3b27c-bf48-4067-b9fa-2312ff79fe66\") " pod="openstack/nova-cell1-db-create-l7mxp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.320484 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wn4xp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.357518 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6e9f-account-create-7788v" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.401952 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01ab5f01-bc77-40d0-aacc-3af0e5c09e3f-operator-scripts\") pod \"nova-cell0-4079-account-create-7bgmp\" (UID: \"01ab5f01-bc77-40d0-aacc-3af0e5c09e3f\") " pod="openstack/nova-cell0-4079-account-create-7bgmp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.402391 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm55t\" (UniqueName: \"kubernetes.io/projected/01ab5f01-bc77-40d0-aacc-3af0e5c09e3f-kube-api-access-gm55t\") pod \"nova-cell0-4079-account-create-7bgmp\" (UID: \"01ab5f01-bc77-40d0-aacc-3af0e5c09e3f\") " pod="openstack/nova-cell0-4079-account-create-7bgmp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.405836 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01ab5f01-bc77-40d0-aacc-3af0e5c09e3f-operator-scripts\") pod \"nova-cell0-4079-account-create-7bgmp\" (UID: \"01ab5f01-bc77-40d0-aacc-3af0e5c09e3f\") " pod="openstack/nova-cell0-4079-account-create-7bgmp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.421130 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-f811-account-create-tt8dk"] Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.421318 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm55t\" (UniqueName: \"kubernetes.io/projected/01ab5f01-bc77-40d0-aacc-3af0e5c09e3f-kube-api-access-gm55t\") pod \"nova-cell0-4079-account-create-7bgmp\" (UID: \"01ab5f01-bc77-40d0-aacc-3af0e5c09e3f\") " pod="openstack/nova-cell0-4079-account-create-7bgmp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.422392 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f811-account-create-tt8dk" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.424482 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.426485 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f811-account-create-tt8dk"] Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.472199 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-l7mxp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.504370 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5cfe52d-a6fb-4bf6-a08b-10f6c6547046-operator-scripts\") pod \"nova-cell1-f811-account-create-tt8dk\" (UID: \"e5cfe52d-a6fb-4bf6-a08b-10f6c6547046\") " pod="openstack/nova-cell1-f811-account-create-tt8dk" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.504689 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvrfx\" (UniqueName: \"kubernetes.io/projected/e5cfe52d-a6fb-4bf6-a08b-10f6c6547046-kube-api-access-dvrfx\") pod \"nova-cell1-f811-account-create-tt8dk\" (UID: \"e5cfe52d-a6fb-4bf6-a08b-10f6c6547046\") " pod="openstack/nova-cell1-f811-account-create-tt8dk" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.603199 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-4079-account-create-7bgmp" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.607160 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvrfx\" (UniqueName: \"kubernetes.io/projected/e5cfe52d-a6fb-4bf6-a08b-10f6c6547046-kube-api-access-dvrfx\") pod \"nova-cell1-f811-account-create-tt8dk\" (UID: \"e5cfe52d-a6fb-4bf6-a08b-10f6c6547046\") " pod="openstack/nova-cell1-f811-account-create-tt8dk" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.607273 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5cfe52d-a6fb-4bf6-a08b-10f6c6547046-operator-scripts\") pod \"nova-cell1-f811-account-create-tt8dk\" (UID: \"e5cfe52d-a6fb-4bf6-a08b-10f6c6547046\") " pod="openstack/nova-cell1-f811-account-create-tt8dk" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.608374 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5cfe52d-a6fb-4bf6-a08b-10f6c6547046-operator-scripts\") pod \"nova-cell1-f811-account-create-tt8dk\" (UID: \"e5cfe52d-a6fb-4bf6-a08b-10f6c6547046\") " pod="openstack/nova-cell1-f811-account-create-tt8dk" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.623295 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvrfx\" (UniqueName: \"kubernetes.io/projected/e5cfe52d-a6fb-4bf6-a08b-10f6c6547046-kube-api-access-dvrfx\") pod \"nova-cell1-f811-account-create-tt8dk\" (UID: \"e5cfe52d-a6fb-4bf6-a08b-10f6c6547046\") " pod="openstack/nova-cell1-f811-account-create-tt8dk" Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.649503 4809 generic.go:334] "Generic (PLEG): container finished" podID="ff37f7ca-34df-422b-98fd-4a63a09c4128" containerID="cdeef0898a6f4fc473a3a31b2881a3fa16c3f2bdd8513e81e2d7fcb1dd6c62fa" exitCode=0 Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.649532 4809 generic.go:334] "Generic (PLEG): container finished" podID="ff37f7ca-34df-422b-98fd-4a63a09c4128" containerID="e7205a02519e1179ec022ddcdf182a82755a35878a71ddf9a3a03b759691ccba" exitCode=2 Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.649540 4809 generic.go:334] "Generic (PLEG): container finished" podID="ff37f7ca-34df-422b-98fd-4a63a09c4128" containerID="3f927e26283f21a361b4680339fb1242ad429849deb9c5a497e2a9f03e3e28d0" exitCode=0 Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.649559 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff37f7ca-34df-422b-98fd-4a63a09c4128","Type":"ContainerDied","Data":"cdeef0898a6f4fc473a3a31b2881a3fa16c3f2bdd8513e81e2d7fcb1dd6c62fa"} Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.649583 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff37f7ca-34df-422b-98fd-4a63a09c4128","Type":"ContainerDied","Data":"e7205a02519e1179ec022ddcdf182a82755a35878a71ddf9a3a03b759691ccba"} Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.649595 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff37f7ca-34df-422b-98fd-4a63a09c4128","Type":"ContainerDied","Data":"3f927e26283f21a361b4680339fb1242ad429849deb9c5a497e2a9f03e3e28d0"} Nov 24 07:12:57 crc kubenswrapper[4809]: I1124 07:12:57.774585 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f811-account-create-tt8dk" Nov 24 07:13:01 crc kubenswrapper[4809]: I1124 07:13:01.705862 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6d74d65d8c-rgmvx" event={"ID":"2c1fe2a7-0b4f-46ee-8368-624499f8c095","Type":"ContainerStarted","Data":"c823c25507a6e65e91660ffde4fa82234a497ea20d104e1287245c1b9725d94a"} Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.128745 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f811-account-create-tt8dk"] Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.135682 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-6e9f-account-create-7788v"] Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.144155 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.148222 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.150948 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-q8dsb"] Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.297150 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-l7mxp"] Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.310323 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-wn4xp"] Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.317060 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-4079-account-create-7bgmp"] Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.455684 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.478516 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.509543 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-config-data\") pod \"ff37f7ca-34df-422b-98fd-4a63a09c4128\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.509672 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-sg-core-conf-yaml\") pod \"ff37f7ca-34df-422b-98fd-4a63a09c4128\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.509761 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-scripts\") pod \"ff37f7ca-34df-422b-98fd-4a63a09c4128\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.509810 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff37f7ca-34df-422b-98fd-4a63a09c4128-run-httpd\") pod \"ff37f7ca-34df-422b-98fd-4a63a09c4128\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.509860 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff37f7ca-34df-422b-98fd-4a63a09c4128-log-httpd\") pod \"ff37f7ca-34df-422b-98fd-4a63a09c4128\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.509888 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-combined-ca-bundle\") pod \"ff37f7ca-34df-422b-98fd-4a63a09c4128\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.509934 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nf2wm\" (UniqueName: \"kubernetes.io/projected/ff37f7ca-34df-422b-98fd-4a63a09c4128-kube-api-access-nf2wm\") pod \"ff37f7ca-34df-422b-98fd-4a63a09c4128\" (UID: \"ff37f7ca-34df-422b-98fd-4a63a09c4128\") " Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.511385 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff37f7ca-34df-422b-98fd-4a63a09c4128-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ff37f7ca-34df-422b-98fd-4a63a09c4128" (UID: "ff37f7ca-34df-422b-98fd-4a63a09c4128"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.511542 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff37f7ca-34df-422b-98fd-4a63a09c4128-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ff37f7ca-34df-422b-98fd-4a63a09c4128" (UID: "ff37f7ca-34df-422b-98fd-4a63a09c4128"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.516254 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-scripts" (OuterVolumeSpecName: "scripts") pod "ff37f7ca-34df-422b-98fd-4a63a09c4128" (UID: "ff37f7ca-34df-422b-98fd-4a63a09c4128"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.533146 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff37f7ca-34df-422b-98fd-4a63a09c4128-kube-api-access-nf2wm" (OuterVolumeSpecName: "kube-api-access-nf2wm") pod "ff37f7ca-34df-422b-98fd-4a63a09c4128" (UID: "ff37f7ca-34df-422b-98fd-4a63a09c4128"). InnerVolumeSpecName "kube-api-access-nf2wm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.542054 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ff37f7ca-34df-422b-98fd-4a63a09c4128" (UID: "ff37f7ca-34df-422b-98fd-4a63a09c4128"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.612410 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nf2wm\" (UniqueName: \"kubernetes.io/projected/ff37f7ca-34df-422b-98fd-4a63a09c4128-kube-api-access-nf2wm\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.612715 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.612724 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.612733 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff37f7ca-34df-422b-98fd-4a63a09c4128-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.612743 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff37f7ca-34df-422b-98fd-4a63a09c4128-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.634414 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff37f7ca-34df-422b-98fd-4a63a09c4128" (UID: "ff37f7ca-34df-422b-98fd-4a63a09c4128"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.656891 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-config-data" (OuterVolumeSpecName: "config-data") pod "ff37f7ca-34df-422b-98fd-4a63a09c4128" (UID: "ff37f7ca-34df-422b-98fd-4a63a09c4128"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.714149 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.714182 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff37f7ca-34df-422b-98fd-4a63a09c4128-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.716508 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-q8dsb" event={"ID":"603e7b5e-01e3-4016-a1bb-9842c9e9ae1e","Type":"ContainerStarted","Data":"932a643312c6abf1ef83683590ac726ed005347389d413ed72583dd8db4815c8"} Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.718219 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-4079-account-create-7bgmp" event={"ID":"01ab5f01-bc77-40d0-aacc-3af0e5c09e3f","Type":"ContainerStarted","Data":"816c89dd38db691c68ad76ddbda702f6d45dfbeb7d9e2c5df63f21b321bfe8e7"} Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.719340 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f811-account-create-tt8dk" event={"ID":"e5cfe52d-a6fb-4bf6-a08b-10f6c6547046","Type":"ContainerStarted","Data":"2641bc83afe90bf78f87135e22d7481428b35852540269ed7ef97546e5f2e2f7"} Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.720323 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6e9f-account-create-7788v" event={"ID":"4197d15d-b7aa-4efe-ad80-c3d731afbc44","Type":"ContainerStarted","Data":"12bd02c8fa17cb0f0a38cb68ec5a429c068874472932784fe21b6dff1a6255f9"} Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.721533 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"45b7caaf-bebc-4bef-804b-a39246b436a0","Type":"ContainerStarted","Data":"9219881d626614a5b938bbcb032f95933d137bb53c2bed4359ccc20208007939"} Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.724128 4809 generic.go:334] "Generic (PLEG): container finished" podID="ff37f7ca-34df-422b-98fd-4a63a09c4128" containerID="1d994e5b49682209744007674d2b4631c46a91caf6c3c6bb5dbe9dcb0f1c4ada" exitCode=0 Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.724190 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff37f7ca-34df-422b-98fd-4a63a09c4128","Type":"ContainerDied","Data":"1d994e5b49682209744007674d2b4631c46a91caf6c3c6bb5dbe9dcb0f1c4ada"} Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.724222 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff37f7ca-34df-422b-98fd-4a63a09c4128","Type":"ContainerDied","Data":"7d06d7b89ff91a4b2108d343a8108c87973718ed7ad3de66bfc5088ee9246fad"} Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.724263 4809 scope.go:117] "RemoveContainer" containerID="cdeef0898a6f4fc473a3a31b2881a3fa16c3f2bdd8513e81e2d7fcb1dd6c62fa" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.724283 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.727275 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-l7mxp" event={"ID":"25b3b27c-bf48-4067-b9fa-2312ff79fe66","Type":"ContainerStarted","Data":"1d1adb1d69d8b99ff36055d83b9d259a5ef899f479a0b0b8b8e5f534a7b5c8fe"} Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.733270 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6d74d65d8c-rgmvx" event={"ID":"2c1fe2a7-0b4f-46ee-8368-624499f8c095","Type":"ContainerStarted","Data":"d863d9483fd17bf1ce562367a48bfafe67d0596918bcedada448e6c1a83c6979"} Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.733311 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6d74d65d8c-rgmvx" event={"ID":"2c1fe2a7-0b4f-46ee-8368-624499f8c095","Type":"ContainerStarted","Data":"e02eddcf7247a8f7647e53d6b1304abbb500d1b9d3fabd822129c4e9cd6c16d2"} Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.735127 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.735190 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.738492 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.733882242 podStartE2EDuration="11.738471691s" podCreationTimestamp="2025-11-24 07:12:51 +0000 UTC" firstStartedPulling="2025-11-24 07:12:52.630476549 +0000 UTC m=+1072.531068244" lastFinishedPulling="2025-11-24 07:13:01.635065988 +0000 UTC m=+1081.535657693" observedRunningTime="2025-11-24 07:13:02.735446491 +0000 UTC m=+1082.636038226" watchObservedRunningTime="2025-11-24 07:13:02.738471691 +0000 UTC m=+1082.639063396" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.740174 4809 generic.go:334] "Generic (PLEG): container finished" podID="1a14aa5c-2f38-4db8-9bf9-422c14699895" containerID="4891063f6682c4b2811a0c696b9c08b9dafd90c63f30e8767f8a00f915cc236a" exitCode=137 Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.740274 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1a14aa5c-2f38-4db8-9bf9-422c14699895","Type":"ContainerDied","Data":"4891063f6682c4b2811a0c696b9c08b9dafd90c63f30e8767f8a00f915cc236a"} Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.740305 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1a14aa5c-2f38-4db8-9bf9-422c14699895","Type":"ContainerDied","Data":"ca8cdde3cb833bbe9a883fccc9869306eb85902a1d8464670d26b39ffbea5d08"} Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.740319 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca8cdde3cb833bbe9a883fccc9869306eb85902a1d8464670d26b39ffbea5d08" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.747358 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wn4xp" event={"ID":"7d43a2b0-ce7c-4392-a104-74b2a2a725c1","Type":"ContainerStarted","Data":"ca2132797ef8ede5539191b3f98d8e1d1d53bccbe2fadd73438e725d3ee64963"} Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.761751 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-6d74d65d8c-rgmvx" podStartSLOduration=7.761729668 podStartE2EDuration="7.761729668s" podCreationTimestamp="2025-11-24 07:12:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:13:02.755477701 +0000 UTC m=+1082.656069396" watchObservedRunningTime="2025-11-24 07:13:02.761729668 +0000 UTC m=+1082.662321363" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.899732 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.912866 4809 scope.go:117] "RemoveContainer" containerID="e7205a02519e1179ec022ddcdf182a82755a35878a71ddf9a3a03b759691ccba" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.917163 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-scripts\") pod \"1a14aa5c-2f38-4db8-9bf9-422c14699895\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.917252 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-config-data\") pod \"1a14aa5c-2f38-4db8-9bf9-422c14699895\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.917344 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-config-data-custom\") pod \"1a14aa5c-2f38-4db8-9bf9-422c14699895\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.917405 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-combined-ca-bundle\") pod \"1a14aa5c-2f38-4db8-9bf9-422c14699895\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.917449 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a14aa5c-2f38-4db8-9bf9-422c14699895-etc-machine-id\") pod \"1a14aa5c-2f38-4db8-9bf9-422c14699895\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.917477 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a14aa5c-2f38-4db8-9bf9-422c14699895-logs\") pod \"1a14aa5c-2f38-4db8-9bf9-422c14699895\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.917500 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df25l\" (UniqueName: \"kubernetes.io/projected/1a14aa5c-2f38-4db8-9bf9-422c14699895-kube-api-access-df25l\") pod \"1a14aa5c-2f38-4db8-9bf9-422c14699895\" (UID: \"1a14aa5c-2f38-4db8-9bf9-422c14699895\") " Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.919387 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a14aa5c-2f38-4db8-9bf9-422c14699895-logs" (OuterVolumeSpecName: "logs") pod "1a14aa5c-2f38-4db8-9bf9-422c14699895" (UID: "1a14aa5c-2f38-4db8-9bf9-422c14699895"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.921245 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1a14aa5c-2f38-4db8-9bf9-422c14699895-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1a14aa5c-2f38-4db8-9bf9-422c14699895" (UID: "1a14aa5c-2f38-4db8-9bf9-422c14699895"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.927457 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-scripts" (OuterVolumeSpecName: "scripts") pod "1a14aa5c-2f38-4db8-9bf9-422c14699895" (UID: "1a14aa5c-2f38-4db8-9bf9-422c14699895"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.935222 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1a14aa5c-2f38-4db8-9bf9-422c14699895" (UID: "1a14aa5c-2f38-4db8-9bf9-422c14699895"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:02 crc kubenswrapper[4809]: I1124 07:13:02.950767 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a14aa5c-2f38-4db8-9bf9-422c14699895-kube-api-access-df25l" (OuterVolumeSpecName: "kube-api-access-df25l") pod "1a14aa5c-2f38-4db8-9bf9-422c14699895" (UID: "1a14aa5c-2f38-4db8-9bf9-422c14699895"). InnerVolumeSpecName "kube-api-access-df25l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.020302 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.020418 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.020487 4809 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a14aa5c-2f38-4db8-9bf9-422c14699895-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.020562 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a14aa5c-2f38-4db8-9bf9-422c14699895-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.020631 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df25l\" (UniqueName: \"kubernetes.io/projected/1a14aa5c-2f38-4db8-9bf9-422c14699895-kube-api-access-df25l\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.035053 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.037622 4809 scope.go:117] "RemoveContainer" containerID="1d994e5b49682209744007674d2b4631c46a91caf6c3c6bb5dbe9dcb0f1c4ada" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.055628 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.083394 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.084915 4809 scope.go:117] "RemoveContainer" containerID="3f927e26283f21a361b4680339fb1242ad429849deb9c5a497e2a9f03e3e28d0" Nov 24 07:13:03 crc kubenswrapper[4809]: E1124 07:13:03.090802 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a14aa5c-2f38-4db8-9bf9-422c14699895" containerName="cinder-api-log" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.091372 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a14aa5c-2f38-4db8-9bf9-422c14699895" containerName="cinder-api-log" Nov 24 07:13:03 crc kubenswrapper[4809]: E1124 07:13:03.091470 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a14aa5c-2f38-4db8-9bf9-422c14699895" containerName="cinder-api" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.091481 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a14aa5c-2f38-4db8-9bf9-422c14699895" containerName="cinder-api" Nov 24 07:13:03 crc kubenswrapper[4809]: E1124 07:13:03.091506 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff37f7ca-34df-422b-98fd-4a63a09c4128" containerName="sg-core" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.091644 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff37f7ca-34df-422b-98fd-4a63a09c4128" containerName="sg-core" Nov 24 07:13:03 crc kubenswrapper[4809]: E1124 07:13:03.091715 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff37f7ca-34df-422b-98fd-4a63a09c4128" containerName="ceilometer-notification-agent" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.091727 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff37f7ca-34df-422b-98fd-4a63a09c4128" containerName="ceilometer-notification-agent" Nov 24 07:13:03 crc kubenswrapper[4809]: E1124 07:13:03.091763 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff37f7ca-34df-422b-98fd-4a63a09c4128" containerName="proxy-httpd" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.092237 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff37f7ca-34df-422b-98fd-4a63a09c4128" containerName="proxy-httpd" Nov 24 07:13:03 crc kubenswrapper[4809]: E1124 07:13:03.092359 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff37f7ca-34df-422b-98fd-4a63a09c4128" containerName="ceilometer-central-agent" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.092391 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff37f7ca-34df-422b-98fd-4a63a09c4128" containerName="ceilometer-central-agent" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.093135 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff37f7ca-34df-422b-98fd-4a63a09c4128" containerName="sg-core" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.093172 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a14aa5c-2f38-4db8-9bf9-422c14699895" containerName="cinder-api" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.093198 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff37f7ca-34df-422b-98fd-4a63a09c4128" containerName="proxy-httpd" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.093209 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff37f7ca-34df-422b-98fd-4a63a09c4128" containerName="ceilometer-central-agent" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.093230 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff37f7ca-34df-422b-98fd-4a63a09c4128" containerName="ceilometer-notification-agent" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.093260 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a14aa5c-2f38-4db8-9bf9-422c14699895" containerName="cinder-api-log" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.111783 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.111910 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.116733 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.119562 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.122212 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6xzv\" (UniqueName: \"kubernetes.io/projected/f62573bb-c652-4198-a70f-aa3a7f49c4f9-kube-api-access-b6xzv\") pod \"ceilometer-0\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " pod="openstack/ceilometer-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.122249 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f62573bb-c652-4198-a70f-aa3a7f49c4f9-run-httpd\") pod \"ceilometer-0\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " pod="openstack/ceilometer-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.122405 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-scripts\") pod \"ceilometer-0\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " pod="openstack/ceilometer-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.122484 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " pod="openstack/ceilometer-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.122547 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " pod="openstack/ceilometer-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.122588 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-config-data\") pod \"ceilometer-0\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " pod="openstack/ceilometer-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.122832 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f62573bb-c652-4198-a70f-aa3a7f49c4f9-log-httpd\") pod \"ceilometer-0\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " pod="openstack/ceilometer-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.129541 4809 scope.go:117] "RemoveContainer" containerID="cdeef0898a6f4fc473a3a31b2881a3fa16c3f2bdd8513e81e2d7fcb1dd6c62fa" Nov 24 07:13:03 crc kubenswrapper[4809]: E1124 07:13:03.130492 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdeef0898a6f4fc473a3a31b2881a3fa16c3f2bdd8513e81e2d7fcb1dd6c62fa\": container with ID starting with cdeef0898a6f4fc473a3a31b2881a3fa16c3f2bdd8513e81e2d7fcb1dd6c62fa not found: ID does not exist" containerID="cdeef0898a6f4fc473a3a31b2881a3fa16c3f2bdd8513e81e2d7fcb1dd6c62fa" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.130548 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdeef0898a6f4fc473a3a31b2881a3fa16c3f2bdd8513e81e2d7fcb1dd6c62fa"} err="failed to get container status \"cdeef0898a6f4fc473a3a31b2881a3fa16c3f2bdd8513e81e2d7fcb1dd6c62fa\": rpc error: code = NotFound desc = could not find container \"cdeef0898a6f4fc473a3a31b2881a3fa16c3f2bdd8513e81e2d7fcb1dd6c62fa\": container with ID starting with cdeef0898a6f4fc473a3a31b2881a3fa16c3f2bdd8513e81e2d7fcb1dd6c62fa not found: ID does not exist" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.130584 4809 scope.go:117] "RemoveContainer" containerID="e7205a02519e1179ec022ddcdf182a82755a35878a71ddf9a3a03b759691ccba" Nov 24 07:13:03 crc kubenswrapper[4809]: E1124 07:13:03.131243 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7205a02519e1179ec022ddcdf182a82755a35878a71ddf9a3a03b759691ccba\": container with ID starting with e7205a02519e1179ec022ddcdf182a82755a35878a71ddf9a3a03b759691ccba not found: ID does not exist" containerID="e7205a02519e1179ec022ddcdf182a82755a35878a71ddf9a3a03b759691ccba" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.131388 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7205a02519e1179ec022ddcdf182a82755a35878a71ddf9a3a03b759691ccba"} err="failed to get container status \"e7205a02519e1179ec022ddcdf182a82755a35878a71ddf9a3a03b759691ccba\": rpc error: code = NotFound desc = could not find container \"e7205a02519e1179ec022ddcdf182a82755a35878a71ddf9a3a03b759691ccba\": container with ID starting with e7205a02519e1179ec022ddcdf182a82755a35878a71ddf9a3a03b759691ccba not found: ID does not exist" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.131411 4809 scope.go:117] "RemoveContainer" containerID="1d994e5b49682209744007674d2b4631c46a91caf6c3c6bb5dbe9dcb0f1c4ada" Nov 24 07:13:03 crc kubenswrapper[4809]: E1124 07:13:03.140322 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d994e5b49682209744007674d2b4631c46a91caf6c3c6bb5dbe9dcb0f1c4ada\": container with ID starting with 1d994e5b49682209744007674d2b4631c46a91caf6c3c6bb5dbe9dcb0f1c4ada not found: ID does not exist" containerID="1d994e5b49682209744007674d2b4631c46a91caf6c3c6bb5dbe9dcb0f1c4ada" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.140390 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d994e5b49682209744007674d2b4631c46a91caf6c3c6bb5dbe9dcb0f1c4ada"} err="failed to get container status \"1d994e5b49682209744007674d2b4631c46a91caf6c3c6bb5dbe9dcb0f1c4ada\": rpc error: code = NotFound desc = could not find container \"1d994e5b49682209744007674d2b4631c46a91caf6c3c6bb5dbe9dcb0f1c4ada\": container with ID starting with 1d994e5b49682209744007674d2b4631c46a91caf6c3c6bb5dbe9dcb0f1c4ada not found: ID does not exist" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.140419 4809 scope.go:117] "RemoveContainer" containerID="3f927e26283f21a361b4680339fb1242ad429849deb9c5a497e2a9f03e3e28d0" Nov 24 07:13:03 crc kubenswrapper[4809]: E1124 07:13:03.144400 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f927e26283f21a361b4680339fb1242ad429849deb9c5a497e2a9f03e3e28d0\": container with ID starting with 3f927e26283f21a361b4680339fb1242ad429849deb9c5a497e2a9f03e3e28d0 not found: ID does not exist" containerID="3f927e26283f21a361b4680339fb1242ad429849deb9c5a497e2a9f03e3e28d0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.144438 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f927e26283f21a361b4680339fb1242ad429849deb9c5a497e2a9f03e3e28d0"} err="failed to get container status \"3f927e26283f21a361b4680339fb1242ad429849deb9c5a497e2a9f03e3e28d0\": rpc error: code = NotFound desc = could not find container \"3f927e26283f21a361b4680339fb1242ad429849deb9c5a497e2a9f03e3e28d0\": container with ID starting with 3f927e26283f21a361b4680339fb1242ad429849deb9c5a497e2a9f03e3e28d0 not found: ID does not exist" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.188340 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a14aa5c-2f38-4db8-9bf9-422c14699895" (UID: "1a14aa5c-2f38-4db8-9bf9-422c14699895"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.205696 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-config-data" (OuterVolumeSpecName: "config-data") pod "1a14aa5c-2f38-4db8-9bf9-422c14699895" (UID: "1a14aa5c-2f38-4db8-9bf9-422c14699895"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.224418 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-scripts\") pod \"ceilometer-0\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " pod="openstack/ceilometer-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.224465 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " pod="openstack/ceilometer-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.224489 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " pod="openstack/ceilometer-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.224511 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-config-data\") pod \"ceilometer-0\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " pod="openstack/ceilometer-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.224591 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f62573bb-c652-4198-a70f-aa3a7f49c4f9-log-httpd\") pod \"ceilometer-0\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " pod="openstack/ceilometer-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.224614 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6xzv\" (UniqueName: \"kubernetes.io/projected/f62573bb-c652-4198-a70f-aa3a7f49c4f9-kube-api-access-b6xzv\") pod \"ceilometer-0\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " pod="openstack/ceilometer-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.224628 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f62573bb-c652-4198-a70f-aa3a7f49c4f9-run-httpd\") pod \"ceilometer-0\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " pod="openstack/ceilometer-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.224681 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.224695 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a14aa5c-2f38-4db8-9bf9-422c14699895-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.225072 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f62573bb-c652-4198-a70f-aa3a7f49c4f9-run-httpd\") pod \"ceilometer-0\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " pod="openstack/ceilometer-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.226203 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f62573bb-c652-4198-a70f-aa3a7f49c4f9-log-httpd\") pod \"ceilometer-0\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " pod="openstack/ceilometer-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.229697 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-scripts\") pod \"ceilometer-0\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " pod="openstack/ceilometer-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.230386 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " pod="openstack/ceilometer-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.230613 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-config-data\") pod \"ceilometer-0\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " pod="openstack/ceilometer-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.231721 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " pod="openstack/ceilometer-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.249689 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6xzv\" (UniqueName: \"kubernetes.io/projected/f62573bb-c652-4198-a70f-aa3a7f49c4f9-kube-api-access-b6xzv\") pod \"ceilometer-0\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " pod="openstack/ceilometer-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.439145 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.816318 4809 generic.go:334] "Generic (PLEG): container finished" podID="7d43a2b0-ce7c-4392-a104-74b2a2a725c1" containerID="d15df5b7d8e85b6a2684a9dbf477405a616065e521d3d10b6c8595965f9577f4" exitCode=0 Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.816583 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wn4xp" event={"ID":"7d43a2b0-ce7c-4392-a104-74b2a2a725c1","Type":"ContainerDied","Data":"d15df5b7d8e85b6a2684a9dbf477405a616065e521d3d10b6c8595965f9577f4"} Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.880667 4809 generic.go:334] "Generic (PLEG): container finished" podID="25b3b27c-bf48-4067-b9fa-2312ff79fe66" containerID="26141a9ca9d9f8e91477625f405b84918fcb3557393e3ba6e068fe5e60c7460f" exitCode=0 Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.881109 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-l7mxp" event={"ID":"25b3b27c-bf48-4067-b9fa-2312ff79fe66","Type":"ContainerDied","Data":"26141a9ca9d9f8e91477625f405b84918fcb3557393e3ba6e068fe5e60c7460f"} Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.914386 4809 generic.go:334] "Generic (PLEG): container finished" podID="603e7b5e-01e3-4016-a1bb-9842c9e9ae1e" containerID="5999e58fde6bd9838c96631ddabea81657a63f5690505976131497160a9369db" exitCode=0 Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.914485 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-q8dsb" event={"ID":"603e7b5e-01e3-4016-a1bb-9842c9e9ae1e","Type":"ContainerDied","Data":"5999e58fde6bd9838c96631ddabea81657a63f5690505976131497160a9369db"} Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.923944 4809 generic.go:334] "Generic (PLEG): container finished" podID="01ab5f01-bc77-40d0-aacc-3af0e5c09e3f" containerID="fe9aa5446cf96b6a630e785504668834e307a395a9d36f2a44ceb7c7db18e1c5" exitCode=0 Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.924108 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-4079-account-create-7bgmp" event={"ID":"01ab5f01-bc77-40d0-aacc-3af0e5c09e3f","Type":"ContainerDied","Data":"fe9aa5446cf96b6a630e785504668834e307a395a9d36f2a44ceb7c7db18e1c5"} Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.929960 4809 generic.go:334] "Generic (PLEG): container finished" podID="e5cfe52d-a6fb-4bf6-a08b-10f6c6547046" containerID="6b7f2215ad35ca949d1beba1c83c3a35b4b828adba3902249bf4be56100f2f4c" exitCode=0 Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.930125 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f811-account-create-tt8dk" event={"ID":"e5cfe52d-a6fb-4bf6-a08b-10f6c6547046","Type":"ContainerDied","Data":"6b7f2215ad35ca949d1beba1c83c3a35b4b828adba3902249bf4be56100f2f4c"} Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.936268 4809 generic.go:334] "Generic (PLEG): container finished" podID="4197d15d-b7aa-4efe-ad80-c3d731afbc44" containerID="e7647fa25d4f30aa0e004b6473877f7a10c062de7bbd7834170076f416f679cf" exitCode=0 Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.936446 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6e9f-account-create-7788v" event={"ID":"4197d15d-b7aa-4efe-ad80-c3d731afbc44","Type":"ContainerDied","Data":"e7647fa25d4f30aa0e004b6473877f7a10c062de7bbd7834170076f416f679cf"} Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.936598 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 07:13:03 crc kubenswrapper[4809]: I1124 07:13:03.995125 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.002128 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.011416 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.044213 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.046351 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.049394 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.049599 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.049732 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.054496 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.170779 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/097f9d2f-9e7a-42b8-a40e-da15a39b146c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.171241 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsjlv\" (UniqueName: \"kubernetes.io/projected/097f9d2f-9e7a-42b8-a40e-da15a39b146c-kube-api-access-xsjlv\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.171331 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/097f9d2f-9e7a-42b8-a40e-da15a39b146c-scripts\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.171377 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/097f9d2f-9e7a-42b8-a40e-da15a39b146c-logs\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.171529 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/097f9d2f-9e7a-42b8-a40e-da15a39b146c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.171719 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/097f9d2f-9e7a-42b8-a40e-da15a39b146c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.171787 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/097f9d2f-9e7a-42b8-a40e-da15a39b146c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.171848 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/097f9d2f-9e7a-42b8-a40e-da15a39b146c-config-data\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.171902 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/097f9d2f-9e7a-42b8-a40e-da15a39b146c-config-data-custom\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.274125 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/097f9d2f-9e7a-42b8-a40e-da15a39b146c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.274199 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/097f9d2f-9e7a-42b8-a40e-da15a39b146c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.274230 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/097f9d2f-9e7a-42b8-a40e-da15a39b146c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.274260 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/097f9d2f-9e7a-42b8-a40e-da15a39b146c-config-data\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.274293 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/097f9d2f-9e7a-42b8-a40e-da15a39b146c-config-data-custom\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.274336 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/097f9d2f-9e7a-42b8-a40e-da15a39b146c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.274430 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/097f9d2f-9e7a-42b8-a40e-da15a39b146c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.274488 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsjlv\" (UniqueName: \"kubernetes.io/projected/097f9d2f-9e7a-42b8-a40e-da15a39b146c-kube-api-access-xsjlv\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.274518 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/097f9d2f-9e7a-42b8-a40e-da15a39b146c-scripts\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.274546 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/097f9d2f-9e7a-42b8-a40e-da15a39b146c-logs\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.275063 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/097f9d2f-9e7a-42b8-a40e-da15a39b146c-logs\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.279844 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/097f9d2f-9e7a-42b8-a40e-da15a39b146c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.280251 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/097f9d2f-9e7a-42b8-a40e-da15a39b146c-config-data-custom\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.289490 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/097f9d2f-9e7a-42b8-a40e-da15a39b146c-scripts\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.289681 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/097f9d2f-9e7a-42b8-a40e-da15a39b146c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.290047 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/097f9d2f-9e7a-42b8-a40e-da15a39b146c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.290212 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/097f9d2f-9e7a-42b8-a40e-da15a39b146c-config-data\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.292344 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsjlv\" (UniqueName: \"kubernetes.io/projected/097f9d2f-9e7a-42b8-a40e-da15a39b146c-kube-api-access-xsjlv\") pod \"cinder-api-0\" (UID: \"097f9d2f-9e7a-42b8-a40e-da15a39b146c\") " pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.378270 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.823707 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.902816 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a14aa5c-2f38-4db8-9bf9-422c14699895" path="/var/lib/kubelet/pods/1a14aa5c-2f38-4db8-9bf9-422c14699895/volumes" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.903844 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff37f7ca-34df-422b-98fd-4a63a09c4128" path="/var/lib/kubelet/pods/ff37f7ca-34df-422b-98fd-4a63a09c4128/volumes" Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.964448 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f62573bb-c652-4198-a70f-aa3a7f49c4f9","Type":"ContainerStarted","Data":"9846ba4af7924c96ce9498edfebbe56896ecb3576883bb4c086c0795d8cbb796"} Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.964723 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f62573bb-c652-4198-a70f-aa3a7f49c4f9","Type":"ContainerStarted","Data":"b3f1cd7411328393aa68cf4c4f58f584d1e681e1783d1d3c715a496ca65ad8a5"} Nov 24 07:13:04 crc kubenswrapper[4809]: I1124 07:13:04.965623 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"097f9d2f-9e7a-42b8-a40e-da15a39b146c","Type":"ContainerStarted","Data":"29d169439bc4bed90144e79c621a3b1df21f64e3cc2a3990c6e5ada16c70795f"} Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.120182 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.161226 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-79cbd445d4-5hkc8" podUID="5b3e7ca6-2814-4a14-b1f4-765f8086927e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.330018 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6e9f-account-create-7788v" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.401710 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4197d15d-b7aa-4efe-ad80-c3d731afbc44-operator-scripts\") pod \"4197d15d-b7aa-4efe-ad80-c3d731afbc44\" (UID: \"4197d15d-b7aa-4efe-ad80-c3d731afbc44\") " Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.401785 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgttz\" (UniqueName: \"kubernetes.io/projected/4197d15d-b7aa-4efe-ad80-c3d731afbc44-kube-api-access-jgttz\") pod \"4197d15d-b7aa-4efe-ad80-c3d731afbc44\" (UID: \"4197d15d-b7aa-4efe-ad80-c3d731afbc44\") " Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.403378 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4197d15d-b7aa-4efe-ad80-c3d731afbc44-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4197d15d-b7aa-4efe-ad80-c3d731afbc44" (UID: "4197d15d-b7aa-4efe-ad80-c3d731afbc44"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.407671 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4197d15d-b7aa-4efe-ad80-c3d731afbc44-kube-api-access-jgttz" (OuterVolumeSpecName: "kube-api-access-jgttz") pod "4197d15d-b7aa-4efe-ad80-c3d731afbc44" (UID: "4197d15d-b7aa-4efe-ad80-c3d731afbc44"). InnerVolumeSpecName "kube-api-access-jgttz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.505070 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4197d15d-b7aa-4efe-ad80-c3d731afbc44-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.505094 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgttz\" (UniqueName: \"kubernetes.io/projected/4197d15d-b7aa-4efe-ad80-c3d731afbc44-kube-api-access-jgttz\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.673577 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-q8dsb" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.679716 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f811-account-create-tt8dk" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.684536 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-4079-account-create-7bgmp" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.696953 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-l7mxp" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.698502 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wn4xp" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.810827 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wk64j\" (UniqueName: \"kubernetes.io/projected/603e7b5e-01e3-4016-a1bb-9842c9e9ae1e-kube-api-access-wk64j\") pod \"603e7b5e-01e3-4016-a1bb-9842c9e9ae1e\" (UID: \"603e7b5e-01e3-4016-a1bb-9842c9e9ae1e\") " Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.811202 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvjk2\" (UniqueName: \"kubernetes.io/projected/25b3b27c-bf48-4067-b9fa-2312ff79fe66-kube-api-access-xvjk2\") pod \"25b3b27c-bf48-4067-b9fa-2312ff79fe66\" (UID: \"25b3b27c-bf48-4067-b9fa-2312ff79fe66\") " Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.811292 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5cfe52d-a6fb-4bf6-a08b-10f6c6547046-operator-scripts\") pod \"e5cfe52d-a6fb-4bf6-a08b-10f6c6547046\" (UID: \"e5cfe52d-a6fb-4bf6-a08b-10f6c6547046\") " Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.811422 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/603e7b5e-01e3-4016-a1bb-9842c9e9ae1e-operator-scripts\") pod \"603e7b5e-01e3-4016-a1bb-9842c9e9ae1e\" (UID: \"603e7b5e-01e3-4016-a1bb-9842c9e9ae1e\") " Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.811467 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gm55t\" (UniqueName: \"kubernetes.io/projected/01ab5f01-bc77-40d0-aacc-3af0e5c09e3f-kube-api-access-gm55t\") pod \"01ab5f01-bc77-40d0-aacc-3af0e5c09e3f\" (UID: \"01ab5f01-bc77-40d0-aacc-3af0e5c09e3f\") " Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.811532 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01ab5f01-bc77-40d0-aacc-3af0e5c09e3f-operator-scripts\") pod \"01ab5f01-bc77-40d0-aacc-3af0e5c09e3f\" (UID: \"01ab5f01-bc77-40d0-aacc-3af0e5c09e3f\") " Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.811610 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d43a2b0-ce7c-4392-a104-74b2a2a725c1-operator-scripts\") pod \"7d43a2b0-ce7c-4392-a104-74b2a2a725c1\" (UID: \"7d43a2b0-ce7c-4392-a104-74b2a2a725c1\") " Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.811686 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68q7k\" (UniqueName: \"kubernetes.io/projected/7d43a2b0-ce7c-4392-a104-74b2a2a725c1-kube-api-access-68q7k\") pod \"7d43a2b0-ce7c-4392-a104-74b2a2a725c1\" (UID: \"7d43a2b0-ce7c-4392-a104-74b2a2a725c1\") " Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.811717 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvrfx\" (UniqueName: \"kubernetes.io/projected/e5cfe52d-a6fb-4bf6-a08b-10f6c6547046-kube-api-access-dvrfx\") pod \"e5cfe52d-a6fb-4bf6-a08b-10f6c6547046\" (UID: \"e5cfe52d-a6fb-4bf6-a08b-10f6c6547046\") " Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.811795 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25b3b27c-bf48-4067-b9fa-2312ff79fe66-operator-scripts\") pod \"25b3b27c-bf48-4067-b9fa-2312ff79fe66\" (UID: \"25b3b27c-bf48-4067-b9fa-2312ff79fe66\") " Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.812604 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/603e7b5e-01e3-4016-a1bb-9842c9e9ae1e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "603e7b5e-01e3-4016-a1bb-9842c9e9ae1e" (UID: "603e7b5e-01e3-4016-a1bb-9842c9e9ae1e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.812690 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25b3b27c-bf48-4067-b9fa-2312ff79fe66-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "25b3b27c-bf48-4067-b9fa-2312ff79fe66" (UID: "25b3b27c-bf48-4067-b9fa-2312ff79fe66"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.813135 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5cfe52d-a6fb-4bf6-a08b-10f6c6547046-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e5cfe52d-a6fb-4bf6-a08b-10f6c6547046" (UID: "e5cfe52d-a6fb-4bf6-a08b-10f6c6547046"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.813822 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d43a2b0-ce7c-4392-a104-74b2a2a725c1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7d43a2b0-ce7c-4392-a104-74b2a2a725c1" (UID: "7d43a2b0-ce7c-4392-a104-74b2a2a725c1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.815089 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab5f01-bc77-40d0-aacc-3af0e5c09e3f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "01ab5f01-bc77-40d0-aacc-3af0e5c09e3f" (UID: "01ab5f01-bc77-40d0-aacc-3af0e5c09e3f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.815549 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25b3b27c-bf48-4067-b9fa-2312ff79fe66-kube-api-access-xvjk2" (OuterVolumeSpecName: "kube-api-access-xvjk2") pod "25b3b27c-bf48-4067-b9fa-2312ff79fe66" (UID: "25b3b27c-bf48-4067-b9fa-2312ff79fe66"). InnerVolumeSpecName "kube-api-access-xvjk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.817827 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5cfe52d-a6fb-4bf6-a08b-10f6c6547046-kube-api-access-dvrfx" (OuterVolumeSpecName: "kube-api-access-dvrfx") pod "e5cfe52d-a6fb-4bf6-a08b-10f6c6547046" (UID: "e5cfe52d-a6fb-4bf6-a08b-10f6c6547046"). InnerVolumeSpecName "kube-api-access-dvrfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.823978 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab5f01-bc77-40d0-aacc-3af0e5c09e3f-kube-api-access-gm55t" (OuterVolumeSpecName: "kube-api-access-gm55t") pod "01ab5f01-bc77-40d0-aacc-3af0e5c09e3f" (UID: "01ab5f01-bc77-40d0-aacc-3af0e5c09e3f"). InnerVolumeSpecName "kube-api-access-gm55t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.828642 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d43a2b0-ce7c-4392-a104-74b2a2a725c1-kube-api-access-68q7k" (OuterVolumeSpecName: "kube-api-access-68q7k") pod "7d43a2b0-ce7c-4392-a104-74b2a2a725c1" (UID: "7d43a2b0-ce7c-4392-a104-74b2a2a725c1"). InnerVolumeSpecName "kube-api-access-68q7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.834151 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/603e7b5e-01e3-4016-a1bb-9842c9e9ae1e-kube-api-access-wk64j" (OuterVolumeSpecName: "kube-api-access-wk64j") pod "603e7b5e-01e3-4016-a1bb-9842c9e9ae1e" (UID: "603e7b5e-01e3-4016-a1bb-9842c9e9ae1e"). InnerVolumeSpecName "kube-api-access-wk64j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.914451 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvjk2\" (UniqueName: \"kubernetes.io/projected/25b3b27c-bf48-4067-b9fa-2312ff79fe66-kube-api-access-xvjk2\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.914479 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5cfe52d-a6fb-4bf6-a08b-10f6c6547046-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.914488 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/603e7b5e-01e3-4016-a1bb-9842c9e9ae1e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.914498 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gm55t\" (UniqueName: \"kubernetes.io/projected/01ab5f01-bc77-40d0-aacc-3af0e5c09e3f-kube-api-access-gm55t\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.914509 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01ab5f01-bc77-40d0-aacc-3af0e5c09e3f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.914521 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d43a2b0-ce7c-4392-a104-74b2a2a725c1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.914532 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68q7k\" (UniqueName: \"kubernetes.io/projected/7d43a2b0-ce7c-4392-a104-74b2a2a725c1-kube-api-access-68q7k\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.914546 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvrfx\" (UniqueName: \"kubernetes.io/projected/e5cfe52d-a6fb-4bf6-a08b-10f6c6547046-kube-api-access-dvrfx\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.914558 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25b3b27c-bf48-4067-b9fa-2312ff79fe66-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.914568 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wk64j\" (UniqueName: \"kubernetes.io/projected/603e7b5e-01e3-4016-a1bb-9842c9e9ae1e-kube-api-access-wk64j\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.991277 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wn4xp" event={"ID":"7d43a2b0-ce7c-4392-a104-74b2a2a725c1","Type":"ContainerDied","Data":"ca2132797ef8ede5539191b3f98d8e1d1d53bccbe2fadd73438e725d3ee64963"} Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.991316 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca2132797ef8ede5539191b3f98d8e1d1d53bccbe2fadd73438e725d3ee64963" Nov 24 07:13:05 crc kubenswrapper[4809]: I1124 07:13:05.991371 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wn4xp" Nov 24 07:13:06 crc kubenswrapper[4809]: I1124 07:13:06.015265 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f62573bb-c652-4198-a70f-aa3a7f49c4f9","Type":"ContainerStarted","Data":"fb51ac869c1bfba24cc5926234151517e60dcc28f7d8663ce9900ff94455e013"} Nov 24 07:13:06 crc kubenswrapper[4809]: I1124 07:13:06.023000 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-l7mxp" event={"ID":"25b3b27c-bf48-4067-b9fa-2312ff79fe66","Type":"ContainerDied","Data":"1d1adb1d69d8b99ff36055d83b9d259a5ef899f479a0b0b8b8e5f534a7b5c8fe"} Nov 24 07:13:06 crc kubenswrapper[4809]: I1124 07:13:06.023117 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d1adb1d69d8b99ff36055d83b9d259a5ef899f479a0b0b8b8e5f534a7b5c8fe" Nov 24 07:13:06 crc kubenswrapper[4809]: I1124 07:13:06.023077 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-l7mxp" Nov 24 07:13:06 crc kubenswrapper[4809]: I1124 07:13:06.025554 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"097f9d2f-9e7a-42b8-a40e-da15a39b146c","Type":"ContainerStarted","Data":"85af317125c2fa210a500552d5a7650212606c6cb777930235e5dc59ce5c60d0"} Nov 24 07:13:06 crc kubenswrapper[4809]: I1124 07:13:06.027368 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-q8dsb" event={"ID":"603e7b5e-01e3-4016-a1bb-9842c9e9ae1e","Type":"ContainerDied","Data":"932a643312c6abf1ef83683590ac726ed005347389d413ed72583dd8db4815c8"} Nov 24 07:13:06 crc kubenswrapper[4809]: I1124 07:13:06.027413 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="932a643312c6abf1ef83683590ac726ed005347389d413ed72583dd8db4815c8" Nov 24 07:13:06 crc kubenswrapper[4809]: I1124 07:13:06.027472 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-q8dsb" Nov 24 07:13:06 crc kubenswrapper[4809]: I1124 07:13:06.037648 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-4079-account-create-7bgmp" event={"ID":"01ab5f01-bc77-40d0-aacc-3af0e5c09e3f","Type":"ContainerDied","Data":"816c89dd38db691c68ad76ddbda702f6d45dfbeb7d9e2c5df63f21b321bfe8e7"} Nov 24 07:13:06 crc kubenswrapper[4809]: I1124 07:13:06.037684 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="816c89dd38db691c68ad76ddbda702f6d45dfbeb7d9e2c5df63f21b321bfe8e7" Nov 24 07:13:06 crc kubenswrapper[4809]: I1124 07:13:06.037731 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-4079-account-create-7bgmp" Nov 24 07:13:06 crc kubenswrapper[4809]: I1124 07:13:06.048159 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f811-account-create-tt8dk" event={"ID":"e5cfe52d-a6fb-4bf6-a08b-10f6c6547046","Type":"ContainerDied","Data":"2641bc83afe90bf78f87135e22d7481428b35852540269ed7ef97546e5f2e2f7"} Nov 24 07:13:06 crc kubenswrapper[4809]: I1124 07:13:06.048221 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2641bc83afe90bf78f87135e22d7481428b35852540269ed7ef97546e5f2e2f7" Nov 24 07:13:06 crc kubenswrapper[4809]: I1124 07:13:06.048320 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f811-account-create-tt8dk" Nov 24 07:13:06 crc kubenswrapper[4809]: I1124 07:13:06.054583 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6e9f-account-create-7788v" event={"ID":"4197d15d-b7aa-4efe-ad80-c3d731afbc44","Type":"ContainerDied","Data":"12bd02c8fa17cb0f0a38cb68ec5a429c068874472932784fe21b6dff1a6255f9"} Nov 24 07:13:06 crc kubenswrapper[4809]: I1124 07:13:06.054622 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12bd02c8fa17cb0f0a38cb68ec5a429c068874472932784fe21b6dff1a6255f9" Nov 24 07:13:06 crc kubenswrapper[4809]: I1124 07:13:06.054634 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6e9f-account-create-7788v" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.067042 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f62573bb-c652-4198-a70f-aa3a7f49c4f9","Type":"ContainerStarted","Data":"bc3edf9b013572c22112e48620bc4820e1474a54f805915685354a91cf07ee52"} Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.070437 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"097f9d2f-9e7a-42b8-a40e-da15a39b146c","Type":"ContainerStarted","Data":"dbba65c0fea3654948cb263eb197c1ff143c14cce7d9cc5736a7caf21ff47c9e"} Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.070787 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.097400 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.097382644 podStartE2EDuration="4.097382644s" podCreationTimestamp="2025-11-24 07:13:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:13:07.091356944 +0000 UTC m=+1086.991948649" watchObservedRunningTime="2025-11-24 07:13:07.097382644 +0000 UTC m=+1086.997974349" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.715260 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-779kx"] Nov 24 07:13:07 crc kubenswrapper[4809]: E1124 07:13:07.716217 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4197d15d-b7aa-4efe-ad80-c3d731afbc44" containerName="mariadb-account-create" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.716254 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4197d15d-b7aa-4efe-ad80-c3d731afbc44" containerName="mariadb-account-create" Nov 24 07:13:07 crc kubenswrapper[4809]: E1124 07:13:07.716274 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d43a2b0-ce7c-4392-a104-74b2a2a725c1" containerName="mariadb-database-create" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.716286 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d43a2b0-ce7c-4392-a104-74b2a2a725c1" containerName="mariadb-database-create" Nov 24 07:13:07 crc kubenswrapper[4809]: E1124 07:13:07.716322 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5cfe52d-a6fb-4bf6-a08b-10f6c6547046" containerName="mariadb-account-create" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.716335 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5cfe52d-a6fb-4bf6-a08b-10f6c6547046" containerName="mariadb-account-create" Nov 24 07:13:07 crc kubenswrapper[4809]: E1124 07:13:07.716358 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="603e7b5e-01e3-4016-a1bb-9842c9e9ae1e" containerName="mariadb-database-create" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.716371 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="603e7b5e-01e3-4016-a1bb-9842c9e9ae1e" containerName="mariadb-database-create" Nov 24 07:13:07 crc kubenswrapper[4809]: E1124 07:13:07.716460 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01ab5f01-bc77-40d0-aacc-3af0e5c09e3f" containerName="mariadb-account-create" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.716474 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="01ab5f01-bc77-40d0-aacc-3af0e5c09e3f" containerName="mariadb-account-create" Nov 24 07:13:07 crc kubenswrapper[4809]: E1124 07:13:07.716492 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25b3b27c-bf48-4067-b9fa-2312ff79fe66" containerName="mariadb-database-create" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.716505 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="25b3b27c-bf48-4067-b9fa-2312ff79fe66" containerName="mariadb-database-create" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.716813 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5cfe52d-a6fb-4bf6-a08b-10f6c6547046" containerName="mariadb-account-create" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.716848 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d43a2b0-ce7c-4392-a104-74b2a2a725c1" containerName="mariadb-database-create" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.716872 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4197d15d-b7aa-4efe-ad80-c3d731afbc44" containerName="mariadb-account-create" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.716891 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="01ab5f01-bc77-40d0-aacc-3af0e5c09e3f" containerName="mariadb-account-create" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.716903 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="603e7b5e-01e3-4016-a1bb-9842c9e9ae1e" containerName="mariadb-database-create" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.716940 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="25b3b27c-bf48-4067-b9fa-2312ff79fe66" containerName="mariadb-database-create" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.717877 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-779kx" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.720306 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-kmj5v" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.723729 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-779kx"] Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.724604 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.724830 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.849940 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6217b72-33d3-4b2e-86a2-10bd847003e4-config-data\") pod \"nova-cell0-conductor-db-sync-779kx\" (UID: \"b6217b72-33d3-4b2e-86a2-10bd847003e4\") " pod="openstack/nova-cell0-conductor-db-sync-779kx" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.850003 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6217b72-33d3-4b2e-86a2-10bd847003e4-scripts\") pod \"nova-cell0-conductor-db-sync-779kx\" (UID: \"b6217b72-33d3-4b2e-86a2-10bd847003e4\") " pod="openstack/nova-cell0-conductor-db-sync-779kx" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.850263 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqwwr\" (UniqueName: \"kubernetes.io/projected/b6217b72-33d3-4b2e-86a2-10bd847003e4-kube-api-access-bqwwr\") pod \"nova-cell0-conductor-db-sync-779kx\" (UID: \"b6217b72-33d3-4b2e-86a2-10bd847003e4\") " pod="openstack/nova-cell0-conductor-db-sync-779kx" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.850480 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6217b72-33d3-4b2e-86a2-10bd847003e4-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-779kx\" (UID: \"b6217b72-33d3-4b2e-86a2-10bd847003e4\") " pod="openstack/nova-cell0-conductor-db-sync-779kx" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.951767 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqwwr\" (UniqueName: \"kubernetes.io/projected/b6217b72-33d3-4b2e-86a2-10bd847003e4-kube-api-access-bqwwr\") pod \"nova-cell0-conductor-db-sync-779kx\" (UID: \"b6217b72-33d3-4b2e-86a2-10bd847003e4\") " pod="openstack/nova-cell0-conductor-db-sync-779kx" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.951877 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6217b72-33d3-4b2e-86a2-10bd847003e4-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-779kx\" (UID: \"b6217b72-33d3-4b2e-86a2-10bd847003e4\") " pod="openstack/nova-cell0-conductor-db-sync-779kx" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.951927 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6217b72-33d3-4b2e-86a2-10bd847003e4-config-data\") pod \"nova-cell0-conductor-db-sync-779kx\" (UID: \"b6217b72-33d3-4b2e-86a2-10bd847003e4\") " pod="openstack/nova-cell0-conductor-db-sync-779kx" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.951950 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6217b72-33d3-4b2e-86a2-10bd847003e4-scripts\") pod \"nova-cell0-conductor-db-sync-779kx\" (UID: \"b6217b72-33d3-4b2e-86a2-10bd847003e4\") " pod="openstack/nova-cell0-conductor-db-sync-779kx" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.957579 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6217b72-33d3-4b2e-86a2-10bd847003e4-config-data\") pod \"nova-cell0-conductor-db-sync-779kx\" (UID: \"b6217b72-33d3-4b2e-86a2-10bd847003e4\") " pod="openstack/nova-cell0-conductor-db-sync-779kx" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.957829 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6217b72-33d3-4b2e-86a2-10bd847003e4-scripts\") pod \"nova-cell0-conductor-db-sync-779kx\" (UID: \"b6217b72-33d3-4b2e-86a2-10bd847003e4\") " pod="openstack/nova-cell0-conductor-db-sync-779kx" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.959493 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6217b72-33d3-4b2e-86a2-10bd847003e4-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-779kx\" (UID: \"b6217b72-33d3-4b2e-86a2-10bd847003e4\") " pod="openstack/nova-cell0-conductor-db-sync-779kx" Nov 24 07:13:07 crc kubenswrapper[4809]: I1124 07:13:07.965750 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqwwr\" (UniqueName: \"kubernetes.io/projected/b6217b72-33d3-4b2e-86a2-10bd847003e4-kube-api-access-bqwwr\") pod \"nova-cell0-conductor-db-sync-779kx\" (UID: \"b6217b72-33d3-4b2e-86a2-10bd847003e4\") " pod="openstack/nova-cell0-conductor-db-sync-779kx" Nov 24 07:13:08 crc kubenswrapper[4809]: I1124 07:13:08.058543 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-779kx" Nov 24 07:13:08 crc kubenswrapper[4809]: I1124 07:13:08.519951 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-779kx"] Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.093768 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-779kx" event={"ID":"b6217b72-33d3-4b2e-86a2-10bd847003e4","Type":"ContainerStarted","Data":"dc12cd3263a05dfe2e57f27f601982352fd2c8c13e510b83f9dc3cb326b4d132"} Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.100280 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f62573bb-c652-4198-a70f-aa3a7f49c4f9","Type":"ContainerStarted","Data":"9236e311367b3c774305960a860fffde731b359e49e545b6fdc90163e930fe96"} Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.100423 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f62573bb-c652-4198-a70f-aa3a7f49c4f9" containerName="ceilometer-central-agent" containerID="cri-o://9846ba4af7924c96ce9498edfebbe56896ecb3576883bb4c086c0795d8cbb796" gracePeriod=30 Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.100471 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.100468 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f62573bb-c652-4198-a70f-aa3a7f49c4f9" containerName="proxy-httpd" containerID="cri-o://9236e311367b3c774305960a860fffde731b359e49e545b6fdc90163e930fe96" gracePeriod=30 Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.100500 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f62573bb-c652-4198-a70f-aa3a7f49c4f9" containerName="sg-core" containerID="cri-o://bc3edf9b013572c22112e48620bc4820e1474a54f805915685354a91cf07ee52" gracePeriod=30 Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.100547 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f62573bb-c652-4198-a70f-aa3a7f49c4f9" containerName="ceilometer-notification-agent" containerID="cri-o://fb51ac869c1bfba24cc5926234151517e60dcc28f7d8663ce9900ff94455e013" gracePeriod=30 Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.121468 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.42105993 podStartE2EDuration="6.121449376s" podCreationTimestamp="2025-11-24 07:13:03 +0000 UTC" firstStartedPulling="2025-11-24 07:13:04.010178546 +0000 UTC m=+1083.910770251" lastFinishedPulling="2025-11-24 07:13:08.710567982 +0000 UTC m=+1088.611159697" observedRunningTime="2025-11-24 07:13:09.119468793 +0000 UTC m=+1089.020060498" watchObservedRunningTime="2025-11-24 07:13:09.121449376 +0000 UTC m=+1089.022041071" Nov 24 07:13:09 crc kubenswrapper[4809]: W1124 07:13:09.263068 4809 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod603e7b5e_01e3_4016_a1bb_9842c9e9ae1e.slice/crio-conmon-5999e58fde6bd9838c96631ddabea81657a63f5690505976131497160a9369db.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod603e7b5e_01e3_4016_a1bb_9842c9e9ae1e.slice/crio-conmon-5999e58fde6bd9838c96631ddabea81657a63f5690505976131497160a9369db.scope: no such file or directory Nov 24 07:13:09 crc kubenswrapper[4809]: W1124 07:13:09.263161 4809 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5cfe52d_a6fb_4bf6_a08b_10f6c6547046.slice/crio-conmon-6b7f2215ad35ca949d1beba1c83c3a35b4b828adba3902249bf4be56100f2f4c.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5cfe52d_a6fb_4bf6_a08b_10f6c6547046.slice/crio-conmon-6b7f2215ad35ca949d1beba1c83c3a35b4b828adba3902249bf4be56100f2f4c.scope: no such file or directory Nov 24 07:13:09 crc kubenswrapper[4809]: W1124 07:13:09.263225 4809 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4197d15d_b7aa_4efe_ad80_c3d731afbc44.slice/crio-e7647fa25d4f30aa0e004b6473877f7a10c062de7bbd7834170076f416f679cf.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4197d15d_b7aa_4efe_ad80_c3d731afbc44.slice/crio-e7647fa25d4f30aa0e004b6473877f7a10c062de7bbd7834170076f416f679cf.scope: no such file or directory Nov 24 07:13:09 crc kubenswrapper[4809]: W1124 07:13:09.263255 4809 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25b3b27c_bf48_4067_b9fa_2312ff79fe66.slice/crio-conmon-26141a9ca9d9f8e91477625f405b84918fcb3557393e3ba6e068fe5e60c7460f.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25b3b27c_bf48_4067_b9fa_2312ff79fe66.slice/crio-conmon-26141a9ca9d9f8e91477625f405b84918fcb3557393e3ba6e068fe5e60c7460f.scope: no such file or directory Nov 24 07:13:09 crc kubenswrapper[4809]: W1124 07:13:09.263279 4809 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d43a2b0_ce7c_4392_a104_74b2a2a725c1.slice/crio-conmon-d15df5b7d8e85b6a2684a9dbf477405a616065e521d3d10b6c8595965f9577f4.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d43a2b0_ce7c_4392_a104_74b2a2a725c1.slice/crio-conmon-d15df5b7d8e85b6a2684a9dbf477405a616065e521d3d10b6c8595965f9577f4.scope: no such file or directory Nov 24 07:13:09 crc kubenswrapper[4809]: W1124 07:13:09.263308 4809 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01ab5f01_bc77_40d0_aacc_3af0e5c09e3f.slice/crio-conmon-fe9aa5446cf96b6a630e785504668834e307a395a9d36f2a44ceb7c7db18e1c5.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01ab5f01_bc77_40d0_aacc_3af0e5c09e3f.slice/crio-conmon-fe9aa5446cf96b6a630e785504668834e307a395a9d36f2a44ceb7c7db18e1c5.scope: no such file or directory Nov 24 07:13:09 crc kubenswrapper[4809]: W1124 07:13:09.263324 4809 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod603e7b5e_01e3_4016_a1bb_9842c9e9ae1e.slice/crio-5999e58fde6bd9838c96631ddabea81657a63f5690505976131497160a9369db.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod603e7b5e_01e3_4016_a1bb_9842c9e9ae1e.slice/crio-5999e58fde6bd9838c96631ddabea81657a63f5690505976131497160a9369db.scope: no such file or directory Nov 24 07:13:09 crc kubenswrapper[4809]: W1124 07:13:09.263339 4809 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5cfe52d_a6fb_4bf6_a08b_10f6c6547046.slice/crio-6b7f2215ad35ca949d1beba1c83c3a35b4b828adba3902249bf4be56100f2f4c.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5cfe52d_a6fb_4bf6_a08b_10f6c6547046.slice/crio-6b7f2215ad35ca949d1beba1c83c3a35b4b828adba3902249bf4be56100f2f4c.scope: no such file or directory Nov 24 07:13:09 crc kubenswrapper[4809]: W1124 07:13:09.263386 4809 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25b3b27c_bf48_4067_b9fa_2312ff79fe66.slice/crio-26141a9ca9d9f8e91477625f405b84918fcb3557393e3ba6e068fe5e60c7460f.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25b3b27c_bf48_4067_b9fa_2312ff79fe66.slice/crio-26141a9ca9d9f8e91477625f405b84918fcb3557393e3ba6e068fe5e60c7460f.scope: no such file or directory Nov 24 07:13:09 crc kubenswrapper[4809]: W1124 07:13:09.263400 4809 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d43a2b0_ce7c_4392_a104_74b2a2a725c1.slice/crio-d15df5b7d8e85b6a2684a9dbf477405a616065e521d3d10b6c8595965f9577f4.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d43a2b0_ce7c_4392_a104_74b2a2a725c1.slice/crio-d15df5b7d8e85b6a2684a9dbf477405a616065e521d3d10b6c8595965f9577f4.scope: no such file or directory Nov 24 07:13:09 crc kubenswrapper[4809]: W1124 07:13:09.263412 4809 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01ab5f01_bc77_40d0_aacc_3af0e5c09e3f.slice/crio-fe9aa5446cf96b6a630e785504668834e307a395a9d36f2a44ceb7c7db18e1c5.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01ab5f01_bc77_40d0_aacc_3af0e5c09e3f.slice/crio-fe9aa5446cf96b6a630e785504668834e307a395a9d36f2a44ceb7c7db18e1c5.scope: no such file or directory Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.593801 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.702358 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b3e7ca6-2814-4a14-b1f4-765f8086927e-horizon-tls-certs\") pod \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.702418 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5b3e7ca6-2814-4a14-b1f4-765f8086927e-config-data\") pod \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.702467 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5b3e7ca6-2814-4a14-b1f4-765f8086927e-scripts\") pod \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.702520 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5b3e7ca6-2814-4a14-b1f4-765f8086927e-horizon-secret-key\") pod \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.702577 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b3e7ca6-2814-4a14-b1f4-765f8086927e-combined-ca-bundle\") pod \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.702665 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbv7j\" (UniqueName: \"kubernetes.io/projected/5b3e7ca6-2814-4a14-b1f4-765f8086927e-kube-api-access-jbv7j\") pod \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.702722 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b3e7ca6-2814-4a14-b1f4-765f8086927e-logs\") pod \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\" (UID: \"5b3e7ca6-2814-4a14-b1f4-765f8086927e\") " Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.703661 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b3e7ca6-2814-4a14-b1f4-765f8086927e-logs" (OuterVolumeSpecName: "logs") pod "5b3e7ca6-2814-4a14-b1f4-765f8086927e" (UID: "5b3e7ca6-2814-4a14-b1f4-765f8086927e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.709101 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b3e7ca6-2814-4a14-b1f4-765f8086927e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "5b3e7ca6-2814-4a14-b1f4-765f8086927e" (UID: "5b3e7ca6-2814-4a14-b1f4-765f8086927e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.712860 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b3e7ca6-2814-4a14-b1f4-765f8086927e-kube-api-access-jbv7j" (OuterVolumeSpecName: "kube-api-access-jbv7j") pod "5b3e7ca6-2814-4a14-b1f4-765f8086927e" (UID: "5b3e7ca6-2814-4a14-b1f4-765f8086927e"). InnerVolumeSpecName "kube-api-access-jbv7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.731372 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b3e7ca6-2814-4a14-b1f4-765f8086927e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b3e7ca6-2814-4a14-b1f4-765f8086927e" (UID: "5b3e7ca6-2814-4a14-b1f4-765f8086927e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.731667 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b3e7ca6-2814-4a14-b1f4-765f8086927e-scripts" (OuterVolumeSpecName: "scripts") pod "5b3e7ca6-2814-4a14-b1f4-765f8086927e" (UID: "5b3e7ca6-2814-4a14-b1f4-765f8086927e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.731810 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b3e7ca6-2814-4a14-b1f4-765f8086927e-config-data" (OuterVolumeSpecName: "config-data") pod "5b3e7ca6-2814-4a14-b1f4-765f8086927e" (UID: "5b3e7ca6-2814-4a14-b1f4-765f8086927e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.754651 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b3e7ca6-2814-4a14-b1f4-765f8086927e-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "5b3e7ca6-2814-4a14-b1f4-765f8086927e" (UID: "5b3e7ca6-2814-4a14-b1f4-765f8086927e"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.805596 4809 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5b3e7ca6-2814-4a14-b1f4-765f8086927e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.805625 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b3e7ca6-2814-4a14-b1f4-765f8086927e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.805636 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbv7j\" (UniqueName: \"kubernetes.io/projected/5b3e7ca6-2814-4a14-b1f4-765f8086927e-kube-api-access-jbv7j\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.805647 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b3e7ca6-2814-4a14-b1f4-765f8086927e-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.805655 4809 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b3e7ca6-2814-4a14-b1f4-765f8086927e-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.805663 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5b3e7ca6-2814-4a14-b1f4-765f8086927e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:09 crc kubenswrapper[4809]: I1124 07:13:09.805672 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5b3e7ca6-2814-4a14-b1f4-765f8086927e-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:10 crc kubenswrapper[4809]: I1124 07:13:10.110456 4809 generic.go:334] "Generic (PLEG): container finished" podID="f62573bb-c652-4198-a70f-aa3a7f49c4f9" containerID="9236e311367b3c774305960a860fffde731b359e49e545b6fdc90163e930fe96" exitCode=0 Nov 24 07:13:10 crc kubenswrapper[4809]: I1124 07:13:10.110488 4809 generic.go:334] "Generic (PLEG): container finished" podID="f62573bb-c652-4198-a70f-aa3a7f49c4f9" containerID="bc3edf9b013572c22112e48620bc4820e1474a54f805915685354a91cf07ee52" exitCode=2 Nov 24 07:13:10 crc kubenswrapper[4809]: I1124 07:13:10.110551 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f62573bb-c652-4198-a70f-aa3a7f49c4f9","Type":"ContainerDied","Data":"9236e311367b3c774305960a860fffde731b359e49e545b6fdc90163e930fe96"} Nov 24 07:13:10 crc kubenswrapper[4809]: I1124 07:13:10.110595 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f62573bb-c652-4198-a70f-aa3a7f49c4f9","Type":"ContainerDied","Data":"bc3edf9b013572c22112e48620bc4820e1474a54f805915685354a91cf07ee52"} Nov 24 07:13:10 crc kubenswrapper[4809]: I1124 07:13:10.110606 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f62573bb-c652-4198-a70f-aa3a7f49c4f9","Type":"ContainerDied","Data":"fb51ac869c1bfba24cc5926234151517e60dcc28f7d8663ce9900ff94455e013"} Nov 24 07:13:10 crc kubenswrapper[4809]: I1124 07:13:10.110572 4809 generic.go:334] "Generic (PLEG): container finished" podID="f62573bb-c652-4198-a70f-aa3a7f49c4f9" containerID="fb51ac869c1bfba24cc5926234151517e60dcc28f7d8663ce9900ff94455e013" exitCode=0 Nov 24 07:13:10 crc kubenswrapper[4809]: I1124 07:13:10.113623 4809 generic.go:334] "Generic (PLEG): container finished" podID="5b3e7ca6-2814-4a14-b1f4-765f8086927e" containerID="1a68297f5bbb4fbec6ca1cf7ed7ec3aaee4d1d4d2c979776b78c71e5790046cc" exitCode=137 Nov 24 07:13:10 crc kubenswrapper[4809]: I1124 07:13:10.113667 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79cbd445d4-5hkc8" Nov 24 07:13:10 crc kubenswrapper[4809]: I1124 07:13:10.113681 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79cbd445d4-5hkc8" event={"ID":"5b3e7ca6-2814-4a14-b1f4-765f8086927e","Type":"ContainerDied","Data":"1a68297f5bbb4fbec6ca1cf7ed7ec3aaee4d1d4d2c979776b78c71e5790046cc"} Nov 24 07:13:10 crc kubenswrapper[4809]: I1124 07:13:10.113709 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79cbd445d4-5hkc8" event={"ID":"5b3e7ca6-2814-4a14-b1f4-765f8086927e","Type":"ContainerDied","Data":"a8a153a7299d3120e11605db7a08ba8650c559c4993bce9f50a9b6d6d11a53cd"} Nov 24 07:13:10 crc kubenswrapper[4809]: I1124 07:13:10.113731 4809 scope.go:117] "RemoveContainer" containerID="54e0622a23539be0abf29e32ee3cbc72076ffc23105f373a61b7e155e5acb13f" Nov 24 07:13:10 crc kubenswrapper[4809]: I1124 07:13:10.152992 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79cbd445d4-5hkc8"] Nov 24 07:13:10 crc kubenswrapper[4809]: I1124 07:13:10.159305 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-79cbd445d4-5hkc8"] Nov 24 07:13:10 crc kubenswrapper[4809]: I1124 07:13:10.293647 4809 scope.go:117] "RemoveContainer" containerID="1a68297f5bbb4fbec6ca1cf7ed7ec3aaee4d1d4d2c979776b78c71e5790046cc" Nov 24 07:13:10 crc kubenswrapper[4809]: I1124 07:13:10.313413 4809 scope.go:117] "RemoveContainer" containerID="54e0622a23539be0abf29e32ee3cbc72076ffc23105f373a61b7e155e5acb13f" Nov 24 07:13:10 crc kubenswrapper[4809]: E1124 07:13:10.313956 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54e0622a23539be0abf29e32ee3cbc72076ffc23105f373a61b7e155e5acb13f\": container with ID starting with 54e0622a23539be0abf29e32ee3cbc72076ffc23105f373a61b7e155e5acb13f not found: ID does not exist" containerID="54e0622a23539be0abf29e32ee3cbc72076ffc23105f373a61b7e155e5acb13f" Nov 24 07:13:10 crc kubenswrapper[4809]: I1124 07:13:10.314020 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54e0622a23539be0abf29e32ee3cbc72076ffc23105f373a61b7e155e5acb13f"} err="failed to get container status \"54e0622a23539be0abf29e32ee3cbc72076ffc23105f373a61b7e155e5acb13f\": rpc error: code = NotFound desc = could not find container \"54e0622a23539be0abf29e32ee3cbc72076ffc23105f373a61b7e155e5acb13f\": container with ID starting with 54e0622a23539be0abf29e32ee3cbc72076ffc23105f373a61b7e155e5acb13f not found: ID does not exist" Nov 24 07:13:10 crc kubenswrapper[4809]: I1124 07:13:10.314067 4809 scope.go:117] "RemoveContainer" containerID="1a68297f5bbb4fbec6ca1cf7ed7ec3aaee4d1d4d2c979776b78c71e5790046cc" Nov 24 07:13:10 crc kubenswrapper[4809]: E1124 07:13:10.314525 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a68297f5bbb4fbec6ca1cf7ed7ec3aaee4d1d4d2c979776b78c71e5790046cc\": container with ID starting with 1a68297f5bbb4fbec6ca1cf7ed7ec3aaee4d1d4d2c979776b78c71e5790046cc not found: ID does not exist" containerID="1a68297f5bbb4fbec6ca1cf7ed7ec3aaee4d1d4d2c979776b78c71e5790046cc" Nov 24 07:13:10 crc kubenswrapper[4809]: I1124 07:13:10.314562 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a68297f5bbb4fbec6ca1cf7ed7ec3aaee4d1d4d2c979776b78c71e5790046cc"} err="failed to get container status \"1a68297f5bbb4fbec6ca1cf7ed7ec3aaee4d1d4d2c979776b78c71e5790046cc\": rpc error: code = NotFound desc = could not find container \"1a68297f5bbb4fbec6ca1cf7ed7ec3aaee4d1d4d2c979776b78c71e5790046cc\": container with ID starting with 1a68297f5bbb4fbec6ca1cf7ed7ec3aaee4d1d4d2c979776b78c71e5790046cc not found: ID does not exist" Nov 24 07:13:10 crc kubenswrapper[4809]: I1124 07:13:10.904429 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b3e7ca6-2814-4a14-b1f4-765f8086927e" path="/var/lib/kubelet/pods/5b3e7ca6-2814-4a14-b1f4-765f8086927e/volumes" Nov 24 07:13:10 crc kubenswrapper[4809]: I1124 07:13:10.984410 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:13:10 crc kubenswrapper[4809]: I1124 07:13:10.985931 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6d74d65d8c-rgmvx" Nov 24 07:13:11 crc kubenswrapper[4809]: E1124 07:13:11.085673 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4197d15d_b7aa_4efe_ad80_c3d731afbc44.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01ab5f01_bc77_40d0_aacc_3af0e5c09e3f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5cfe52d_a6fb_4bf6_a08b_10f6c6547046.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5cfe52d_a6fb_4bf6_a08b_10f6c6547046.slice/crio-2641bc83afe90bf78f87135e22d7481428b35852540269ed7ef97546e5f2e2f7\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01ab5f01_bc77_40d0_aacc_3af0e5c09e3f.slice/crio-816c89dd38db691c68ad76ddbda702f6d45dfbeb7d9e2c5df63f21b321bfe8e7\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b3e7ca6_2814_4a14_b1f4_765f8086927e.slice/crio-1a68297f5bbb4fbec6ca1cf7ed7ec3aaee4d1d4d2c979776b78c71e5790046cc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4197d15d_b7aa_4efe_ad80_c3d731afbc44.slice/crio-12bd02c8fa17cb0f0a38cb68ec5a429c068874472932784fe21b6dff1a6255f9\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25b3b27c_bf48_4067_b9fa_2312ff79fe66.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b3e7ca6_2814_4a14_b1f4_765f8086927e.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d43a2b0_ce7c_4392_a104_74b2a2a725c1.slice/crio-ca2132797ef8ede5539191b3f98d8e1d1d53bccbe2fadd73438e725d3ee64963\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod603e7b5e_01e3_4016_a1bb_9842c9e9ae1e.slice/crio-932a643312c6abf1ef83683590ac726ed005347389d413ed72583dd8db4815c8\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b3e7ca6_2814_4a14_b1f4_765f8086927e.slice/crio-a8a153a7299d3120e11605db7a08ba8650c559c4993bce9f50a9b6d6d11a53cd\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod603e7b5e_01e3_4016_a1bb_9842c9e9ae1e.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d43a2b0_ce7c_4392_a104_74b2a2a725c1.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b3e7ca6_2814_4a14_b1f4_765f8086927e.slice/crio-conmon-1a68297f5bbb4fbec6ca1cf7ed7ec3aaee4d1d4d2c979776b78c71e5790046cc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25b3b27c_bf48_4067_b9fa_2312ff79fe66.slice/crio-1d1adb1d69d8b99ff36055d83b9d259a5ef899f479a0b0b8b8e5f534a7b5c8fe\": RecentStats: unable to find data in memory cache]" Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.154019 4809 generic.go:334] "Generic (PLEG): container finished" podID="f62573bb-c652-4198-a70f-aa3a7f49c4f9" containerID="9846ba4af7924c96ce9498edfebbe56896ecb3576883bb4c086c0795d8cbb796" exitCode=0 Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.154800 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f62573bb-c652-4198-a70f-aa3a7f49c4f9","Type":"ContainerDied","Data":"9846ba4af7924c96ce9498edfebbe56896ecb3576883bb4c086c0795d8cbb796"} Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.640050 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.738383 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-combined-ca-bundle\") pod \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.738473 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f62573bb-c652-4198-a70f-aa3a7f49c4f9-log-httpd\") pod \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.738499 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-config-data\") pod \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.738532 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f62573bb-c652-4198-a70f-aa3a7f49c4f9-run-httpd\") pod \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.738559 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-sg-core-conf-yaml\") pod \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.738632 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-scripts\") pod \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.738657 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6xzv\" (UniqueName: \"kubernetes.io/projected/f62573bb-c652-4198-a70f-aa3a7f49c4f9-kube-api-access-b6xzv\") pod \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\" (UID: \"f62573bb-c652-4198-a70f-aa3a7f49c4f9\") " Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.740157 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f62573bb-c652-4198-a70f-aa3a7f49c4f9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f62573bb-c652-4198-a70f-aa3a7f49c4f9" (UID: "f62573bb-c652-4198-a70f-aa3a7f49c4f9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.741052 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f62573bb-c652-4198-a70f-aa3a7f49c4f9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f62573bb-c652-4198-a70f-aa3a7f49c4f9" (UID: "f62573bb-c652-4198-a70f-aa3a7f49c4f9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.744504 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-scripts" (OuterVolumeSpecName: "scripts") pod "f62573bb-c652-4198-a70f-aa3a7f49c4f9" (UID: "f62573bb-c652-4198-a70f-aa3a7f49c4f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.744685 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f62573bb-c652-4198-a70f-aa3a7f49c4f9-kube-api-access-b6xzv" (OuterVolumeSpecName: "kube-api-access-b6xzv") pod "f62573bb-c652-4198-a70f-aa3a7f49c4f9" (UID: "f62573bb-c652-4198-a70f-aa3a7f49c4f9"). InnerVolumeSpecName "kube-api-access-b6xzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.780476 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f62573bb-c652-4198-a70f-aa3a7f49c4f9" (UID: "f62573bb-c652-4198-a70f-aa3a7f49c4f9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.826915 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f62573bb-c652-4198-a70f-aa3a7f49c4f9" (UID: "f62573bb-c652-4198-a70f-aa3a7f49c4f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.840391 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f62573bb-c652-4198-a70f-aa3a7f49c4f9-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.840421 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f62573bb-c652-4198-a70f-aa3a7f49c4f9-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.840430 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.840440 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.840448 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6xzv\" (UniqueName: \"kubernetes.io/projected/f62573bb-c652-4198-a70f-aa3a7f49c4f9-kube-api-access-b6xzv\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.840457 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.849147 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-config-data" (OuterVolumeSpecName: "config-data") pod "f62573bb-c652-4198-a70f-aa3a7f49c4f9" (UID: "f62573bb-c652-4198-a70f-aa3a7f49c4f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:11 crc kubenswrapper[4809]: I1124 07:13:11.941980 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f62573bb-c652-4198-a70f-aa3a7f49c4f9-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.169245 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f62573bb-c652-4198-a70f-aa3a7f49c4f9","Type":"ContainerDied","Data":"b3f1cd7411328393aa68cf4c4f58f584d1e681e1783d1d3c715a496ca65ad8a5"} Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.169290 4809 scope.go:117] "RemoveContainer" containerID="9236e311367b3c774305960a860fffde731b359e49e545b6fdc90163e930fe96" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.169405 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.218891 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.227649 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.254133 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:13:12 crc kubenswrapper[4809]: E1124 07:13:12.254463 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f62573bb-c652-4198-a70f-aa3a7f49c4f9" containerName="ceilometer-central-agent" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.254478 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f62573bb-c652-4198-a70f-aa3a7f49c4f9" containerName="ceilometer-central-agent" Nov 24 07:13:12 crc kubenswrapper[4809]: E1124 07:13:12.254493 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f62573bb-c652-4198-a70f-aa3a7f49c4f9" containerName="proxy-httpd" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.254500 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f62573bb-c652-4198-a70f-aa3a7f49c4f9" containerName="proxy-httpd" Nov 24 07:13:12 crc kubenswrapper[4809]: E1124 07:13:12.254522 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f62573bb-c652-4198-a70f-aa3a7f49c4f9" containerName="sg-core" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.254528 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f62573bb-c652-4198-a70f-aa3a7f49c4f9" containerName="sg-core" Nov 24 07:13:12 crc kubenswrapper[4809]: E1124 07:13:12.254540 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b3e7ca6-2814-4a14-b1f4-765f8086927e" containerName="horizon" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.254546 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b3e7ca6-2814-4a14-b1f4-765f8086927e" containerName="horizon" Nov 24 07:13:12 crc kubenswrapper[4809]: E1124 07:13:12.254558 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f62573bb-c652-4198-a70f-aa3a7f49c4f9" containerName="ceilometer-notification-agent" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.254564 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f62573bb-c652-4198-a70f-aa3a7f49c4f9" containerName="ceilometer-notification-agent" Nov 24 07:13:12 crc kubenswrapper[4809]: E1124 07:13:12.254587 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b3e7ca6-2814-4a14-b1f4-765f8086927e" containerName="horizon-log" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.254594 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b3e7ca6-2814-4a14-b1f4-765f8086927e" containerName="horizon-log" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.254760 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f62573bb-c652-4198-a70f-aa3a7f49c4f9" containerName="ceilometer-notification-agent" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.254771 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f62573bb-c652-4198-a70f-aa3a7f49c4f9" containerName="ceilometer-central-agent" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.254781 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b3e7ca6-2814-4a14-b1f4-765f8086927e" containerName="horizon-log" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.254793 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b3e7ca6-2814-4a14-b1f4-765f8086927e" containerName="horizon" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.254802 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f62573bb-c652-4198-a70f-aa3a7f49c4f9" containerName="sg-core" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.254814 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f62573bb-c652-4198-a70f-aa3a7f49c4f9" containerName="proxy-httpd" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.256571 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.259570 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.259793 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.274649 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.348995 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q978h\" (UniqueName: \"kubernetes.io/projected/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-kube-api-access-q978h\") pod \"ceilometer-0\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.349068 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-log-httpd\") pod \"ceilometer-0\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.349095 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.349132 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-run-httpd\") pod \"ceilometer-0\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.349189 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-config-data\") pod \"ceilometer-0\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.349205 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.349237 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-scripts\") pod \"ceilometer-0\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.450914 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-run-httpd\") pod \"ceilometer-0\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.451019 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-config-data\") pod \"ceilometer-0\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.451036 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.451074 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-scripts\") pod \"ceilometer-0\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.451107 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q978h\" (UniqueName: \"kubernetes.io/projected/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-kube-api-access-q978h\") pod \"ceilometer-0\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.451200 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-log-httpd\") pod \"ceilometer-0\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.451229 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.451594 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-run-httpd\") pod \"ceilometer-0\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.451708 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-log-httpd\") pod \"ceilometer-0\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.455026 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.455530 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.457675 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-scripts\") pod \"ceilometer-0\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.459152 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-config-data\") pod \"ceilometer-0\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.472724 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q978h\" (UniqueName: \"kubernetes.io/projected/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-kube-api-access-q978h\") pod \"ceilometer-0\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.591472 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:13:12 crc kubenswrapper[4809]: I1124 07:13:12.903778 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f62573bb-c652-4198-a70f-aa3a7f49c4f9" path="/var/lib/kubelet/pods/f62573bb-c652-4198-a70f-aa3a7f49c4f9/volumes" Nov 24 07:13:16 crc kubenswrapper[4809]: I1124 07:13:16.226612 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 24 07:13:16 crc kubenswrapper[4809]: I1124 07:13:16.419158 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:13:16 crc kubenswrapper[4809]: I1124 07:13:16.419784 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="29842a6a-e532-4757-829e-418ecc738eca" containerName="glance-httpd" containerID="cri-o://abb9fd0265dd261d032eb959dd6b9eb41c5e9d55f981ba63a0a7413c61c584a9" gracePeriod=30 Nov 24 07:13:16 crc kubenswrapper[4809]: I1124 07:13:16.419683 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="29842a6a-e532-4757-829e-418ecc738eca" containerName="glance-log" containerID="cri-o://d606986fa98397e571fbd9ac327ccc96cd4c2550e7f041c4a3383d845df59ee0" gracePeriod=30 Nov 24 07:13:16 crc kubenswrapper[4809]: I1124 07:13:16.794777 4809 scope.go:117] "RemoveContainer" containerID="bc3edf9b013572c22112e48620bc4820e1474a54f805915685354a91cf07ee52" Nov 24 07:13:16 crc kubenswrapper[4809]: I1124 07:13:16.849345 4809 scope.go:117] "RemoveContainer" containerID="fb51ac869c1bfba24cc5926234151517e60dcc28f7d8663ce9900ff94455e013" Nov 24 07:13:17 crc kubenswrapper[4809]: I1124 07:13:17.031591 4809 scope.go:117] "RemoveContainer" containerID="9846ba4af7924c96ce9498edfebbe56896ecb3576883bb4c086c0795d8cbb796" Nov 24 07:13:17 crc kubenswrapper[4809]: I1124 07:13:17.229319 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-779kx" event={"ID":"b6217b72-33d3-4b2e-86a2-10bd847003e4","Type":"ContainerStarted","Data":"c02fa30371d4010318a10fd542ad941354969f99e6d49d8655b2cc7d516dc5ed"} Nov 24 07:13:17 crc kubenswrapper[4809]: I1124 07:13:17.232411 4809 generic.go:334] "Generic (PLEG): container finished" podID="29842a6a-e532-4757-829e-418ecc738eca" containerID="d606986fa98397e571fbd9ac327ccc96cd4c2550e7f041c4a3383d845df59ee0" exitCode=143 Nov 24 07:13:17 crc kubenswrapper[4809]: I1124 07:13:17.232480 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"29842a6a-e532-4757-829e-418ecc738eca","Type":"ContainerDied","Data":"d606986fa98397e571fbd9ac327ccc96cd4c2550e7f041c4a3383d845df59ee0"} Nov 24 07:13:17 crc kubenswrapper[4809]: I1124 07:13:17.249167 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-779kx" podStartSLOduration=1.900690103 podStartE2EDuration="10.249150486s" podCreationTimestamp="2025-11-24 07:13:07 +0000 UTC" firstStartedPulling="2025-11-24 07:13:08.527419997 +0000 UTC m=+1088.428011702" lastFinishedPulling="2025-11-24 07:13:16.87588037 +0000 UTC m=+1096.776472085" observedRunningTime="2025-11-24 07:13:17.244727639 +0000 UTC m=+1097.145319354" watchObservedRunningTime="2025-11-24 07:13:17.249150486 +0000 UTC m=+1097.149742191" Nov 24 07:13:17 crc kubenswrapper[4809]: I1124 07:13:17.282570 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:13:18 crc kubenswrapper[4809]: I1124 07:13:18.043470 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:13:18 crc kubenswrapper[4809]: I1124 07:13:18.043788 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:13:18 crc kubenswrapper[4809]: I1124 07:13:18.246462 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de9cfa7f-1c42-4acc-bca3-d750e80ca2df","Type":"ContainerStarted","Data":"7060ce936d4610777f74f46ed7964a43b70beb8ebcf75f3020cfa4aa292a7ad5"} Nov 24 07:13:18 crc kubenswrapper[4809]: I1124 07:13:18.246542 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de9cfa7f-1c42-4acc-bca3-d750e80ca2df","Type":"ContainerStarted","Data":"f9039d5269c457fd102d52e1a0e883dd3d205cb34f812bffe12865c40cbd594c"} Nov 24 07:13:19 crc kubenswrapper[4809]: I1124 07:13:19.260145 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de9cfa7f-1c42-4acc-bca3-d750e80ca2df","Type":"ContainerStarted","Data":"e6f54d113f3f9d93dd7bdfb375edfad35f8e6e9c84c8f280e99705da719f0ba5"} Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.047982 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.203044 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-config-data\") pod \"29842a6a-e532-4757-829e-418ecc738eca\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.203095 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"29842a6a-e532-4757-829e-418ecc738eca\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.203153 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-combined-ca-bundle\") pod \"29842a6a-e532-4757-829e-418ecc738eca\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.203188 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-public-tls-certs\") pod \"29842a6a-e532-4757-829e-418ecc738eca\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.203223 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29842a6a-e532-4757-829e-418ecc738eca-logs\") pod \"29842a6a-e532-4757-829e-418ecc738eca\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.203254 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgzbl\" (UniqueName: \"kubernetes.io/projected/29842a6a-e532-4757-829e-418ecc738eca-kube-api-access-pgzbl\") pod \"29842a6a-e532-4757-829e-418ecc738eca\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.203268 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-scripts\") pod \"29842a6a-e532-4757-829e-418ecc738eca\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.203290 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/29842a6a-e532-4757-829e-418ecc738eca-httpd-run\") pod \"29842a6a-e532-4757-829e-418ecc738eca\" (UID: \"29842a6a-e532-4757-829e-418ecc738eca\") " Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.204407 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29842a6a-e532-4757-829e-418ecc738eca-logs" (OuterVolumeSpecName: "logs") pod "29842a6a-e532-4757-829e-418ecc738eca" (UID: "29842a6a-e532-4757-829e-418ecc738eca"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.204552 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29842a6a-e532-4757-829e-418ecc738eca-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "29842a6a-e532-4757-829e-418ecc738eca" (UID: "29842a6a-e532-4757-829e-418ecc738eca"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.209256 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29842a6a-e532-4757-829e-418ecc738eca-kube-api-access-pgzbl" (OuterVolumeSpecName: "kube-api-access-pgzbl") pod "29842a6a-e532-4757-829e-418ecc738eca" (UID: "29842a6a-e532-4757-829e-418ecc738eca"). InnerVolumeSpecName "kube-api-access-pgzbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.209922 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "29842a6a-e532-4757-829e-418ecc738eca" (UID: "29842a6a-e532-4757-829e-418ecc738eca"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.216174 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-scripts" (OuterVolumeSpecName: "scripts") pod "29842a6a-e532-4757-829e-418ecc738eca" (UID: "29842a6a-e532-4757-829e-418ecc738eca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.237646 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "29842a6a-e532-4757-829e-418ecc738eca" (UID: "29842a6a-e532-4757-829e-418ecc738eca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.261527 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "29842a6a-e532-4757-829e-418ecc738eca" (UID: "29842a6a-e532-4757-829e-418ecc738eca"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.275215 4809 generic.go:334] "Generic (PLEG): container finished" podID="29842a6a-e532-4757-829e-418ecc738eca" containerID="abb9fd0265dd261d032eb959dd6b9eb41c5e9d55f981ba63a0a7413c61c584a9" exitCode=0 Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.275266 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.275286 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"29842a6a-e532-4757-829e-418ecc738eca","Type":"ContainerDied","Data":"abb9fd0265dd261d032eb959dd6b9eb41c5e9d55f981ba63a0a7413c61c584a9"} Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.276158 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"29842a6a-e532-4757-829e-418ecc738eca","Type":"ContainerDied","Data":"dbba0e81948331aafcc04809174f2e9a3efe2000e5fbbc136048d5e9435da9a5"} Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.276192 4809 scope.go:117] "RemoveContainer" containerID="abb9fd0265dd261d032eb959dd6b9eb41c5e9d55f981ba63a0a7413c61c584a9" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.279236 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de9cfa7f-1c42-4acc-bca3-d750e80ca2df","Type":"ContainerStarted","Data":"c1e846d6eae83fd73b20ddfce4036215bb537449030aabe05da8ec6fc4250115"} Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.285464 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-config-data" (OuterVolumeSpecName: "config-data") pod "29842a6a-e532-4757-829e-418ecc738eca" (UID: "29842a6a-e532-4757-829e-418ecc738eca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.305705 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.305924 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.305999 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.306056 4809 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.306107 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29842a6a-e532-4757-829e-418ecc738eca-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.306163 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgzbl\" (UniqueName: \"kubernetes.io/projected/29842a6a-e532-4757-829e-418ecc738eca-kube-api-access-pgzbl\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.306227 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29842a6a-e532-4757-829e-418ecc738eca-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.306313 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/29842a6a-e532-4757-829e-418ecc738eca-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.336274 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.346871 4809 scope.go:117] "RemoveContainer" containerID="d606986fa98397e571fbd9ac327ccc96cd4c2550e7f041c4a3383d845df59ee0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.375811 4809 scope.go:117] "RemoveContainer" containerID="abb9fd0265dd261d032eb959dd6b9eb41c5e9d55f981ba63a0a7413c61c584a9" Nov 24 07:13:20 crc kubenswrapper[4809]: E1124 07:13:20.376258 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abb9fd0265dd261d032eb959dd6b9eb41c5e9d55f981ba63a0a7413c61c584a9\": container with ID starting with abb9fd0265dd261d032eb959dd6b9eb41c5e9d55f981ba63a0a7413c61c584a9 not found: ID does not exist" containerID="abb9fd0265dd261d032eb959dd6b9eb41c5e9d55f981ba63a0a7413c61c584a9" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.376292 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abb9fd0265dd261d032eb959dd6b9eb41c5e9d55f981ba63a0a7413c61c584a9"} err="failed to get container status \"abb9fd0265dd261d032eb959dd6b9eb41c5e9d55f981ba63a0a7413c61c584a9\": rpc error: code = NotFound desc = could not find container \"abb9fd0265dd261d032eb959dd6b9eb41c5e9d55f981ba63a0a7413c61c584a9\": container with ID starting with abb9fd0265dd261d032eb959dd6b9eb41c5e9d55f981ba63a0a7413c61c584a9 not found: ID does not exist" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.376311 4809 scope.go:117] "RemoveContainer" containerID="d606986fa98397e571fbd9ac327ccc96cd4c2550e7f041c4a3383d845df59ee0" Nov 24 07:13:20 crc kubenswrapper[4809]: E1124 07:13:20.376609 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d606986fa98397e571fbd9ac327ccc96cd4c2550e7f041c4a3383d845df59ee0\": container with ID starting with d606986fa98397e571fbd9ac327ccc96cd4c2550e7f041c4a3383d845df59ee0 not found: ID does not exist" containerID="d606986fa98397e571fbd9ac327ccc96cd4c2550e7f041c4a3383d845df59ee0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.376660 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d606986fa98397e571fbd9ac327ccc96cd4c2550e7f041c4a3383d845df59ee0"} err="failed to get container status \"d606986fa98397e571fbd9ac327ccc96cd4c2550e7f041c4a3383d845df59ee0\": rpc error: code = NotFound desc = could not find container \"d606986fa98397e571fbd9ac327ccc96cd4c2550e7f041c4a3383d845df59ee0\": container with ID starting with d606986fa98397e571fbd9ac327ccc96cd4c2550e7f041c4a3383d845df59ee0 not found: ID does not exist" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.407756 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.609827 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.625594 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.636596 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:13:20 crc kubenswrapper[4809]: E1124 07:13:20.636998 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29842a6a-e532-4757-829e-418ecc738eca" containerName="glance-httpd" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.637014 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="29842a6a-e532-4757-829e-418ecc738eca" containerName="glance-httpd" Nov 24 07:13:20 crc kubenswrapper[4809]: E1124 07:13:20.637032 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29842a6a-e532-4757-829e-418ecc738eca" containerName="glance-log" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.637039 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="29842a6a-e532-4757-829e-418ecc738eca" containerName="glance-log" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.637238 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="29842a6a-e532-4757-829e-418ecc738eca" containerName="glance-httpd" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.637254 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="29842a6a-e532-4757-829e-418ecc738eca" containerName="glance-log" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.638124 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.641856 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.642066 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.649913 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.712564 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfb0ea66-5650-4f85-a36a-06b60949a4bf-scripts\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.712661 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7s7r\" (UniqueName: \"kubernetes.io/projected/cfb0ea66-5650-4f85-a36a-06b60949a4bf-kube-api-access-l7s7r\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.712695 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfb0ea66-5650-4f85-a36a-06b60949a4bf-config-data\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.712747 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cfb0ea66-5650-4f85-a36a-06b60949a4bf-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.712846 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfb0ea66-5650-4f85-a36a-06b60949a4bf-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.712873 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfb0ea66-5650-4f85-a36a-06b60949a4bf-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.712915 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfb0ea66-5650-4f85-a36a-06b60949a4bf-logs\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.712940 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.814322 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cfb0ea66-5650-4f85-a36a-06b60949a4bf-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.814453 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfb0ea66-5650-4f85-a36a-06b60949a4bf-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.814481 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfb0ea66-5650-4f85-a36a-06b60949a4bf-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.814528 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.814551 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfb0ea66-5650-4f85-a36a-06b60949a4bf-logs\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.814574 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfb0ea66-5650-4f85-a36a-06b60949a4bf-scripts\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.814645 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7s7r\" (UniqueName: \"kubernetes.io/projected/cfb0ea66-5650-4f85-a36a-06b60949a4bf-kube-api-access-l7s7r\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.814681 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfb0ea66-5650-4f85-a36a-06b60949a4bf-config-data\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.815903 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.820479 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cfb0ea66-5650-4f85-a36a-06b60949a4bf-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.820918 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfb0ea66-5650-4f85-a36a-06b60949a4bf-logs\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.822609 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfb0ea66-5650-4f85-a36a-06b60949a4bf-config-data\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.823239 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfb0ea66-5650-4f85-a36a-06b60949a4bf-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.824927 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfb0ea66-5650-4f85-a36a-06b60949a4bf-scripts\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.828521 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfb0ea66-5650-4f85-a36a-06b60949a4bf-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.835708 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7s7r\" (UniqueName: \"kubernetes.io/projected/cfb0ea66-5650-4f85-a36a-06b60949a4bf-kube-api-access-l7s7r\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.858884 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"cfb0ea66-5650-4f85-a36a-06b60949a4bf\") " pod="openstack/glance-default-external-api-0" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.866903 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.867181 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0bf846c6-e717-4fab-8fe0-0d43d1747ce4" containerName="glance-log" containerID="cri-o://b15a6958eff3edbf8366b4ad13629375e82ded4b11e08ed984f470b7add61f0e" gracePeriod=30 Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.867663 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0bf846c6-e717-4fab-8fe0-0d43d1747ce4" containerName="glance-httpd" containerID="cri-o://ecc506aaa274fb6334ab278f177ffc8f8b8f2706a9ac605330d06ff15b1cd040" gracePeriod=30 Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.910231 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29842a6a-e532-4757-829e-418ecc738eca" path="/var/lib/kubelet/pods/29842a6a-e532-4757-829e-418ecc738eca/volumes" Nov 24 07:13:20 crc kubenswrapper[4809]: I1124 07:13:20.975446 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 07:13:21 crc kubenswrapper[4809]: I1124 07:13:21.323493 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de9cfa7f-1c42-4acc-bca3-d750e80ca2df","Type":"ContainerStarted","Data":"f5f9965b8316317d7f110490a6332a0509cf7d39f16b9dff6d66b6739156771a"} Nov 24 07:13:21 crc kubenswrapper[4809]: I1124 07:13:21.323827 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 07:13:21 crc kubenswrapper[4809]: I1124 07:13:21.327308 4809 generic.go:334] "Generic (PLEG): container finished" podID="0bf846c6-e717-4fab-8fe0-0d43d1747ce4" containerID="b15a6958eff3edbf8366b4ad13629375e82ded4b11e08ed984f470b7add61f0e" exitCode=143 Nov 24 07:13:21 crc kubenswrapper[4809]: I1124 07:13:21.327350 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0bf846c6-e717-4fab-8fe0-0d43d1747ce4","Type":"ContainerDied","Data":"b15a6958eff3edbf8366b4ad13629375e82ded4b11e08ed984f470b7add61f0e"} Nov 24 07:13:21 crc kubenswrapper[4809]: I1124 07:13:21.346436 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=6.121888375 podStartE2EDuration="9.346421173s" podCreationTimestamp="2025-11-24 07:13:12 +0000 UTC" firstStartedPulling="2025-11-24 07:13:17.289485486 +0000 UTC m=+1097.190077211" lastFinishedPulling="2025-11-24 07:13:20.514018284 +0000 UTC m=+1100.414610009" observedRunningTime="2025-11-24 07:13:21.34405364 +0000 UTC m=+1101.244645345" watchObservedRunningTime="2025-11-24 07:13:21.346421173 +0000 UTC m=+1101.247012878" Nov 24 07:13:21 crc kubenswrapper[4809]: I1124 07:13:21.518011 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:13:21 crc kubenswrapper[4809]: W1124 07:13:21.575960 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfb0ea66_5650_4f85_a36a_06b60949a4bf.slice/crio-3968f77c86705be7a6abe461dadcb0ce974bc3f033ab435a83182888bea3c99b WatchSource:0}: Error finding container 3968f77c86705be7a6abe461dadcb0ce974bc3f033ab435a83182888bea3c99b: Status 404 returned error can't find the container with id 3968f77c86705be7a6abe461dadcb0ce974bc3f033ab435a83182888bea3c99b Nov 24 07:13:22 crc kubenswrapper[4809]: I1124 07:13:22.350274 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cfb0ea66-5650-4f85-a36a-06b60949a4bf","Type":"ContainerStarted","Data":"5adb36960be5a66f3d911248c51bfda63525111142aab94acae0959574914f2b"} Nov 24 07:13:22 crc kubenswrapper[4809]: I1124 07:13:22.350800 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cfb0ea66-5650-4f85-a36a-06b60949a4bf","Type":"ContainerStarted","Data":"3968f77c86705be7a6abe461dadcb0ce974bc3f033ab435a83182888bea3c99b"} Nov 24 07:13:22 crc kubenswrapper[4809]: I1124 07:13:22.382476 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:13:23 crc kubenswrapper[4809]: I1124 07:13:23.359586 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cfb0ea66-5650-4f85-a36a-06b60949a4bf","Type":"ContainerStarted","Data":"faf3501187138fffa5288f797a849220197d5cda3cb56cee406aa3d20b48d082"} Nov 24 07:13:23 crc kubenswrapper[4809]: I1124 07:13:23.359738 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de9cfa7f-1c42-4acc-bca3-d750e80ca2df" containerName="ceilometer-central-agent" containerID="cri-o://7060ce936d4610777f74f46ed7964a43b70beb8ebcf75f3020cfa4aa292a7ad5" gracePeriod=30 Nov 24 07:13:23 crc kubenswrapper[4809]: I1124 07:13:23.359810 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de9cfa7f-1c42-4acc-bca3-d750e80ca2df" containerName="sg-core" containerID="cri-o://c1e846d6eae83fd73b20ddfce4036215bb537449030aabe05da8ec6fc4250115" gracePeriod=30 Nov 24 07:13:23 crc kubenswrapper[4809]: I1124 07:13:23.359833 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de9cfa7f-1c42-4acc-bca3-d750e80ca2df" containerName="proxy-httpd" containerID="cri-o://f5f9965b8316317d7f110490a6332a0509cf7d39f16b9dff6d66b6739156771a" gracePeriod=30 Nov 24 07:13:23 crc kubenswrapper[4809]: I1124 07:13:23.359827 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de9cfa7f-1c42-4acc-bca3-d750e80ca2df" containerName="ceilometer-notification-agent" containerID="cri-o://e6f54d113f3f9d93dd7bdfb375edfad35f8e6e9c84c8f280e99705da719f0ba5" gracePeriod=30 Nov 24 07:13:23 crc kubenswrapper[4809]: I1124 07:13:23.386645 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.386620022 podStartE2EDuration="3.386620022s" podCreationTimestamp="2025-11-24 07:13:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:13:23.385797881 +0000 UTC m=+1103.286389586" watchObservedRunningTime="2025-11-24 07:13:23.386620022 +0000 UTC m=+1103.287211727" Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.395171 4809 generic.go:334] "Generic (PLEG): container finished" podID="0bf846c6-e717-4fab-8fe0-0d43d1747ce4" containerID="ecc506aaa274fb6334ab278f177ffc8f8b8f2706a9ac605330d06ff15b1cd040" exitCode=0 Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.395259 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0bf846c6-e717-4fab-8fe0-0d43d1747ce4","Type":"ContainerDied","Data":"ecc506aaa274fb6334ab278f177ffc8f8b8f2706a9ac605330d06ff15b1cd040"} Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.398756 4809 generic.go:334] "Generic (PLEG): container finished" podID="de9cfa7f-1c42-4acc-bca3-d750e80ca2df" containerID="f5f9965b8316317d7f110490a6332a0509cf7d39f16b9dff6d66b6739156771a" exitCode=0 Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.398786 4809 generic.go:334] "Generic (PLEG): container finished" podID="de9cfa7f-1c42-4acc-bca3-d750e80ca2df" containerID="c1e846d6eae83fd73b20ddfce4036215bb537449030aabe05da8ec6fc4250115" exitCode=2 Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.398798 4809 generic.go:334] "Generic (PLEG): container finished" podID="de9cfa7f-1c42-4acc-bca3-d750e80ca2df" containerID="e6f54d113f3f9d93dd7bdfb375edfad35f8e6e9c84c8f280e99705da719f0ba5" exitCode=0 Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.399004 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de9cfa7f-1c42-4acc-bca3-d750e80ca2df","Type":"ContainerDied","Data":"f5f9965b8316317d7f110490a6332a0509cf7d39f16b9dff6d66b6739156771a"} Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.399029 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de9cfa7f-1c42-4acc-bca3-d750e80ca2df","Type":"ContainerDied","Data":"c1e846d6eae83fd73b20ddfce4036215bb537449030aabe05da8ec6fc4250115"} Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.399040 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de9cfa7f-1c42-4acc-bca3-d750e80ca2df","Type":"ContainerDied","Data":"e6f54d113f3f9d93dd7bdfb375edfad35f8e6e9c84c8f280e99705da719f0ba5"} Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.536278 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.630607 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-internal-tls-certs\") pod \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.630697 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xl578\" (UniqueName: \"kubernetes.io/projected/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-kube-api-access-xl578\") pod \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.630752 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.630787 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-httpd-run\") pod \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.630805 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-logs\") pod \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.630842 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-combined-ca-bundle\") pod \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.630903 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-config-data\") pod \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.630928 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-scripts\") pod \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\" (UID: \"0bf846c6-e717-4fab-8fe0-0d43d1747ce4\") " Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.631764 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0bf846c6-e717-4fab-8fe0-0d43d1747ce4" (UID: "0bf846c6-e717-4fab-8fe0-0d43d1747ce4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.631806 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-logs" (OuterVolumeSpecName: "logs") pod "0bf846c6-e717-4fab-8fe0-0d43d1747ce4" (UID: "0bf846c6-e717-4fab-8fe0-0d43d1747ce4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.636587 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-kube-api-access-xl578" (OuterVolumeSpecName: "kube-api-access-xl578") pod "0bf846c6-e717-4fab-8fe0-0d43d1747ce4" (UID: "0bf846c6-e717-4fab-8fe0-0d43d1747ce4"). InnerVolumeSpecName "kube-api-access-xl578". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.641327 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-scripts" (OuterVolumeSpecName: "scripts") pod "0bf846c6-e717-4fab-8fe0-0d43d1747ce4" (UID: "0bf846c6-e717-4fab-8fe0-0d43d1747ce4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.645467 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "0bf846c6-e717-4fab-8fe0-0d43d1747ce4" (UID: "0bf846c6-e717-4fab-8fe0-0d43d1747ce4"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.680434 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0bf846c6-e717-4fab-8fe0-0d43d1747ce4" (UID: "0bf846c6-e717-4fab-8fe0-0d43d1747ce4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.688576 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0bf846c6-e717-4fab-8fe0-0d43d1747ce4" (UID: "0bf846c6-e717-4fab-8fe0-0d43d1747ce4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.700157 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-config-data" (OuterVolumeSpecName: "config-data") pod "0bf846c6-e717-4fab-8fe0-0d43d1747ce4" (UID: "0bf846c6-e717-4fab-8fe0-0d43d1747ce4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.733091 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.733126 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.733136 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.733148 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.733158 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.733168 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.733175 4809 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.733185 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xl578\" (UniqueName: \"kubernetes.io/projected/0bf846c6-e717-4fab-8fe0-0d43d1747ce4-kube-api-access-xl578\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.755893 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 24 07:13:24 crc kubenswrapper[4809]: I1124 07:13:24.834352 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.409567 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0bf846c6-e717-4fab-8fe0-0d43d1747ce4","Type":"ContainerDied","Data":"f82b340d02e56520c08a4ed546872cc9bdce4cf3ada4b594d0987107ca2bc1bf"} Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.409615 4809 scope.go:117] "RemoveContainer" containerID="ecc506aaa274fb6334ab278f177ffc8f8b8f2706a9ac605330d06ff15b1cd040" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.409793 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.436110 4809 scope.go:117] "RemoveContainer" containerID="b15a6958eff3edbf8366b4ad13629375e82ded4b11e08ed984f470b7add61f0e" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.444728 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.463927 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.476909 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:13:25 crc kubenswrapper[4809]: E1124 07:13:25.477324 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bf846c6-e717-4fab-8fe0-0d43d1747ce4" containerName="glance-httpd" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.477342 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bf846c6-e717-4fab-8fe0-0d43d1747ce4" containerName="glance-httpd" Nov 24 07:13:25 crc kubenswrapper[4809]: E1124 07:13:25.477373 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bf846c6-e717-4fab-8fe0-0d43d1747ce4" containerName="glance-log" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.477381 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bf846c6-e717-4fab-8fe0-0d43d1747ce4" containerName="glance-log" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.477563 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bf846c6-e717-4fab-8fe0-0d43d1747ce4" containerName="glance-log" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.477597 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bf846c6-e717-4fab-8fe0-0d43d1747ce4" containerName="glance-httpd" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.478752 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.485680 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.486127 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.491990 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.545627 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.545690 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lmvb\" (UniqueName: \"kubernetes.io/projected/91d79979-625d-4ad3-bbb5-f3f35e7142a4-kube-api-access-6lmvb\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.545727 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91d79979-625d-4ad3-bbb5-f3f35e7142a4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.545808 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/91d79979-625d-4ad3-bbb5-f3f35e7142a4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.545831 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91d79979-625d-4ad3-bbb5-f3f35e7142a4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.545886 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91d79979-625d-4ad3-bbb5-f3f35e7142a4-logs\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.545912 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91d79979-625d-4ad3-bbb5-f3f35e7142a4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.545939 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/91d79979-625d-4ad3-bbb5-f3f35e7142a4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.647776 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/91d79979-625d-4ad3-bbb5-f3f35e7142a4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.648061 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91d79979-625d-4ad3-bbb5-f3f35e7142a4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.648082 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91d79979-625d-4ad3-bbb5-f3f35e7142a4-logs\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.648107 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91d79979-625d-4ad3-bbb5-f3f35e7142a4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.648128 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/91d79979-625d-4ad3-bbb5-f3f35e7142a4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.648202 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.648234 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lmvb\" (UniqueName: \"kubernetes.io/projected/91d79979-625d-4ad3-bbb5-f3f35e7142a4-kube-api-access-6lmvb\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.648258 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91d79979-625d-4ad3-bbb5-f3f35e7142a4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.648354 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/91d79979-625d-4ad3-bbb5-f3f35e7142a4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.648539 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91d79979-625d-4ad3-bbb5-f3f35e7142a4-logs\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.648760 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.652434 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/91d79979-625d-4ad3-bbb5-f3f35e7142a4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.652548 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91d79979-625d-4ad3-bbb5-f3f35e7142a4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.657721 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91d79979-625d-4ad3-bbb5-f3f35e7142a4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.667533 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91d79979-625d-4ad3-bbb5-f3f35e7142a4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.671737 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lmvb\" (UniqueName: \"kubernetes.io/projected/91d79979-625d-4ad3-bbb5-f3f35e7142a4-kube-api-access-6lmvb\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.676078 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"91d79979-625d-4ad3-bbb5-f3f35e7142a4\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:13:25 crc kubenswrapper[4809]: I1124 07:13:25.803576 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 07:13:26 crc kubenswrapper[4809]: I1124 07:13:26.325880 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:13:26 crc kubenswrapper[4809]: W1124 07:13:26.330686 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91d79979_625d_4ad3_bbb5_f3f35e7142a4.slice/crio-be79c842992b64dd68ce652b05af161c7d7128a2869e171e585f947c8148920e WatchSource:0}: Error finding container be79c842992b64dd68ce652b05af161c7d7128a2869e171e585f947c8148920e: Status 404 returned error can't find the container with id be79c842992b64dd68ce652b05af161c7d7128a2869e171e585f947c8148920e Nov 24 07:13:26 crc kubenswrapper[4809]: I1124 07:13:26.429738 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"91d79979-625d-4ad3-bbb5-f3f35e7142a4","Type":"ContainerStarted","Data":"be79c842992b64dd68ce652b05af161c7d7128a2869e171e585f947c8148920e"} Nov 24 07:13:26 crc kubenswrapper[4809]: I1124 07:13:26.860701 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:13:26 crc kubenswrapper[4809]: I1124 07:13:26.911302 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bf846c6-e717-4fab-8fe0-0d43d1747ce4" path="/var/lib/kubelet/pods/0bf846c6-e717-4fab-8fe0-0d43d1747ce4/volumes" Nov 24 07:13:26 crc kubenswrapper[4809]: I1124 07:13:26.976353 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-run-httpd\") pod \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " Nov 24 07:13:26 crc kubenswrapper[4809]: I1124 07:13:26.976437 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-config-data\") pod \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " Nov 24 07:13:26 crc kubenswrapper[4809]: I1124 07:13:26.976578 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-combined-ca-bundle\") pod \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " Nov 24 07:13:26 crc kubenswrapper[4809]: I1124 07:13:26.976616 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-sg-core-conf-yaml\") pod \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " Nov 24 07:13:26 crc kubenswrapper[4809]: I1124 07:13:26.976732 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-log-httpd\") pod \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " Nov 24 07:13:26 crc kubenswrapper[4809]: I1124 07:13:26.976778 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-scripts\") pod \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " Nov 24 07:13:26 crc kubenswrapper[4809]: I1124 07:13:26.976800 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q978h\" (UniqueName: \"kubernetes.io/projected/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-kube-api-access-q978h\") pod \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\" (UID: \"de9cfa7f-1c42-4acc-bca3-d750e80ca2df\") " Nov 24 07:13:26 crc kubenswrapper[4809]: I1124 07:13:26.978228 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "de9cfa7f-1c42-4acc-bca3-d750e80ca2df" (UID: "de9cfa7f-1c42-4acc-bca3-d750e80ca2df"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:13:26 crc kubenswrapper[4809]: I1124 07:13:26.978682 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "de9cfa7f-1c42-4acc-bca3-d750e80ca2df" (UID: "de9cfa7f-1c42-4acc-bca3-d750e80ca2df"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:13:26 crc kubenswrapper[4809]: I1124 07:13:26.981941 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-kube-api-access-q978h" (OuterVolumeSpecName: "kube-api-access-q978h") pod "de9cfa7f-1c42-4acc-bca3-d750e80ca2df" (UID: "de9cfa7f-1c42-4acc-bca3-d750e80ca2df"). InnerVolumeSpecName "kube-api-access-q978h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:13:26 crc kubenswrapper[4809]: I1124 07:13:26.983997 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-scripts" (OuterVolumeSpecName: "scripts") pod "de9cfa7f-1c42-4acc-bca3-d750e80ca2df" (UID: "de9cfa7f-1c42-4acc-bca3-d750e80ca2df"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.016441 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "de9cfa7f-1c42-4acc-bca3-d750e80ca2df" (UID: "de9cfa7f-1c42-4acc-bca3-d750e80ca2df"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.058001 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de9cfa7f-1c42-4acc-bca3-d750e80ca2df" (UID: "de9cfa7f-1c42-4acc-bca3-d750e80ca2df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.079567 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.079600 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.079609 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.079618 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.079626 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q978h\" (UniqueName: \"kubernetes.io/projected/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-kube-api-access-q978h\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.079639 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.095414 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-config-data" (OuterVolumeSpecName: "config-data") pod "de9cfa7f-1c42-4acc-bca3-d750e80ca2df" (UID: "de9cfa7f-1c42-4acc-bca3-d750e80ca2df"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.181394 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de9cfa7f-1c42-4acc-bca3-d750e80ca2df-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.444023 4809 generic.go:334] "Generic (PLEG): container finished" podID="de9cfa7f-1c42-4acc-bca3-d750e80ca2df" containerID="7060ce936d4610777f74f46ed7964a43b70beb8ebcf75f3020cfa4aa292a7ad5" exitCode=0 Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.444086 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.444110 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de9cfa7f-1c42-4acc-bca3-d750e80ca2df","Type":"ContainerDied","Data":"7060ce936d4610777f74f46ed7964a43b70beb8ebcf75f3020cfa4aa292a7ad5"} Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.444157 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de9cfa7f-1c42-4acc-bca3-d750e80ca2df","Type":"ContainerDied","Data":"f9039d5269c457fd102d52e1a0e883dd3d205cb34f812bffe12865c40cbd594c"} Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.444176 4809 scope.go:117] "RemoveContainer" containerID="f5f9965b8316317d7f110490a6332a0509cf7d39f16b9dff6d66b6739156771a" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.454320 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"91d79979-625d-4ad3-bbb5-f3f35e7142a4","Type":"ContainerStarted","Data":"c81a01c3139f56b14614459eb43ccb292c2d913bc0f5e603fa0e4ca45e8dfd2f"} Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.464583 4809 scope.go:117] "RemoveContainer" containerID="c1e846d6eae83fd73b20ddfce4036215bb537449030aabe05da8ec6fc4250115" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.490450 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.490432882 podStartE2EDuration="2.490432882s" podCreationTimestamp="2025-11-24 07:13:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:13:27.477655363 +0000 UTC m=+1107.378247068" watchObservedRunningTime="2025-11-24 07:13:27.490432882 +0000 UTC m=+1107.391024587" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.497408 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.504778 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.527472 4809 scope.go:117] "RemoveContainer" containerID="e6f54d113f3f9d93dd7bdfb375edfad35f8e6e9c84c8f280e99705da719f0ba5" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.527691 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:13:27 crc kubenswrapper[4809]: E1124 07:13:27.528273 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de9cfa7f-1c42-4acc-bca3-d750e80ca2df" containerName="sg-core" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.528299 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="de9cfa7f-1c42-4acc-bca3-d750e80ca2df" containerName="sg-core" Nov 24 07:13:27 crc kubenswrapper[4809]: E1124 07:13:27.528310 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de9cfa7f-1c42-4acc-bca3-d750e80ca2df" containerName="proxy-httpd" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.528318 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="de9cfa7f-1c42-4acc-bca3-d750e80ca2df" containerName="proxy-httpd" Nov 24 07:13:27 crc kubenswrapper[4809]: E1124 07:13:27.528361 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de9cfa7f-1c42-4acc-bca3-d750e80ca2df" containerName="ceilometer-notification-agent" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.528370 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="de9cfa7f-1c42-4acc-bca3-d750e80ca2df" containerName="ceilometer-notification-agent" Nov 24 07:13:27 crc kubenswrapper[4809]: E1124 07:13:27.528387 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de9cfa7f-1c42-4acc-bca3-d750e80ca2df" containerName="ceilometer-central-agent" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.528394 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="de9cfa7f-1c42-4acc-bca3-d750e80ca2df" containerName="ceilometer-central-agent" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.532114 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="de9cfa7f-1c42-4acc-bca3-d750e80ca2df" containerName="ceilometer-central-agent" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.532182 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="de9cfa7f-1c42-4acc-bca3-d750e80ca2df" containerName="proxy-httpd" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.532197 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="de9cfa7f-1c42-4acc-bca3-d750e80ca2df" containerName="sg-core" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.532208 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="de9cfa7f-1c42-4acc-bca3-d750e80ca2df" containerName="ceilometer-notification-agent" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.538093 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.539784 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.540590 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.540724 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.589319 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llw4x\" (UniqueName: \"kubernetes.io/projected/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-kube-api-access-llw4x\") pod \"ceilometer-0\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " pod="openstack/ceilometer-0" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.589375 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-scripts\") pod \"ceilometer-0\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " pod="openstack/ceilometer-0" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.589728 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-log-httpd\") pod \"ceilometer-0\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " pod="openstack/ceilometer-0" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.589782 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " pod="openstack/ceilometer-0" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.589895 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-run-httpd\") pod \"ceilometer-0\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " pod="openstack/ceilometer-0" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.590884 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-config-data\") pod \"ceilometer-0\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " pod="openstack/ceilometer-0" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.591108 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " pod="openstack/ceilometer-0" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.593372 4809 scope.go:117] "RemoveContainer" containerID="7060ce936d4610777f74f46ed7964a43b70beb8ebcf75f3020cfa4aa292a7ad5" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.627028 4809 scope.go:117] "RemoveContainer" containerID="f5f9965b8316317d7f110490a6332a0509cf7d39f16b9dff6d66b6739156771a" Nov 24 07:13:27 crc kubenswrapper[4809]: E1124 07:13:27.627454 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5f9965b8316317d7f110490a6332a0509cf7d39f16b9dff6d66b6739156771a\": container with ID starting with f5f9965b8316317d7f110490a6332a0509cf7d39f16b9dff6d66b6739156771a not found: ID does not exist" containerID="f5f9965b8316317d7f110490a6332a0509cf7d39f16b9dff6d66b6739156771a" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.627504 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5f9965b8316317d7f110490a6332a0509cf7d39f16b9dff6d66b6739156771a"} err="failed to get container status \"f5f9965b8316317d7f110490a6332a0509cf7d39f16b9dff6d66b6739156771a\": rpc error: code = NotFound desc = could not find container \"f5f9965b8316317d7f110490a6332a0509cf7d39f16b9dff6d66b6739156771a\": container with ID starting with f5f9965b8316317d7f110490a6332a0509cf7d39f16b9dff6d66b6739156771a not found: ID does not exist" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.627541 4809 scope.go:117] "RemoveContainer" containerID="c1e846d6eae83fd73b20ddfce4036215bb537449030aabe05da8ec6fc4250115" Nov 24 07:13:27 crc kubenswrapper[4809]: E1124 07:13:27.627857 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1e846d6eae83fd73b20ddfce4036215bb537449030aabe05da8ec6fc4250115\": container with ID starting with c1e846d6eae83fd73b20ddfce4036215bb537449030aabe05da8ec6fc4250115 not found: ID does not exist" containerID="c1e846d6eae83fd73b20ddfce4036215bb537449030aabe05da8ec6fc4250115" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.627882 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1e846d6eae83fd73b20ddfce4036215bb537449030aabe05da8ec6fc4250115"} err="failed to get container status \"c1e846d6eae83fd73b20ddfce4036215bb537449030aabe05da8ec6fc4250115\": rpc error: code = NotFound desc = could not find container \"c1e846d6eae83fd73b20ddfce4036215bb537449030aabe05da8ec6fc4250115\": container with ID starting with c1e846d6eae83fd73b20ddfce4036215bb537449030aabe05da8ec6fc4250115 not found: ID does not exist" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.627901 4809 scope.go:117] "RemoveContainer" containerID="e6f54d113f3f9d93dd7bdfb375edfad35f8e6e9c84c8f280e99705da719f0ba5" Nov 24 07:13:27 crc kubenswrapper[4809]: E1124 07:13:27.628209 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6f54d113f3f9d93dd7bdfb375edfad35f8e6e9c84c8f280e99705da719f0ba5\": container with ID starting with e6f54d113f3f9d93dd7bdfb375edfad35f8e6e9c84c8f280e99705da719f0ba5 not found: ID does not exist" containerID="e6f54d113f3f9d93dd7bdfb375edfad35f8e6e9c84c8f280e99705da719f0ba5" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.628230 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6f54d113f3f9d93dd7bdfb375edfad35f8e6e9c84c8f280e99705da719f0ba5"} err="failed to get container status \"e6f54d113f3f9d93dd7bdfb375edfad35f8e6e9c84c8f280e99705da719f0ba5\": rpc error: code = NotFound desc = could not find container \"e6f54d113f3f9d93dd7bdfb375edfad35f8e6e9c84c8f280e99705da719f0ba5\": container with ID starting with e6f54d113f3f9d93dd7bdfb375edfad35f8e6e9c84c8f280e99705da719f0ba5 not found: ID does not exist" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.628247 4809 scope.go:117] "RemoveContainer" containerID="7060ce936d4610777f74f46ed7964a43b70beb8ebcf75f3020cfa4aa292a7ad5" Nov 24 07:13:27 crc kubenswrapper[4809]: E1124 07:13:27.628447 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7060ce936d4610777f74f46ed7964a43b70beb8ebcf75f3020cfa4aa292a7ad5\": container with ID starting with 7060ce936d4610777f74f46ed7964a43b70beb8ebcf75f3020cfa4aa292a7ad5 not found: ID does not exist" containerID="7060ce936d4610777f74f46ed7964a43b70beb8ebcf75f3020cfa4aa292a7ad5" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.628469 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7060ce936d4610777f74f46ed7964a43b70beb8ebcf75f3020cfa4aa292a7ad5"} err="failed to get container status \"7060ce936d4610777f74f46ed7964a43b70beb8ebcf75f3020cfa4aa292a7ad5\": rpc error: code = NotFound desc = could not find container \"7060ce936d4610777f74f46ed7964a43b70beb8ebcf75f3020cfa4aa292a7ad5\": container with ID starting with 7060ce936d4610777f74f46ed7964a43b70beb8ebcf75f3020cfa4aa292a7ad5 not found: ID does not exist" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.693011 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llw4x\" (UniqueName: \"kubernetes.io/projected/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-kube-api-access-llw4x\") pod \"ceilometer-0\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " pod="openstack/ceilometer-0" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.693061 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-scripts\") pod \"ceilometer-0\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " pod="openstack/ceilometer-0" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.693107 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-log-httpd\") pod \"ceilometer-0\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " pod="openstack/ceilometer-0" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.693123 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " pod="openstack/ceilometer-0" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.693186 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-run-httpd\") pod \"ceilometer-0\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " pod="openstack/ceilometer-0" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.693252 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-config-data\") pod \"ceilometer-0\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " pod="openstack/ceilometer-0" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.693320 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " pod="openstack/ceilometer-0" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.693612 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-log-httpd\") pod \"ceilometer-0\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " pod="openstack/ceilometer-0" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.693722 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-run-httpd\") pod \"ceilometer-0\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " pod="openstack/ceilometer-0" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.698639 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " pod="openstack/ceilometer-0" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.700293 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " pod="openstack/ceilometer-0" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.700334 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-config-data\") pod \"ceilometer-0\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " pod="openstack/ceilometer-0" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.702128 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-scripts\") pod \"ceilometer-0\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " pod="openstack/ceilometer-0" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.711938 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llw4x\" (UniqueName: \"kubernetes.io/projected/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-kube-api-access-llw4x\") pod \"ceilometer-0\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " pod="openstack/ceilometer-0" Nov 24 07:13:27 crc kubenswrapper[4809]: I1124 07:13:27.895694 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:13:28 crc kubenswrapper[4809]: I1124 07:13:28.328685 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:13:28 crc kubenswrapper[4809]: W1124 07:13:28.330908 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4aa2fc56_9abd_4a5a_ac78_8ad5ccec8e12.slice/crio-19539bd948969f6ad4f75c9beef8a27054770fe8039f316a8335bc5f1fa6fc1d WatchSource:0}: Error finding container 19539bd948969f6ad4f75c9beef8a27054770fe8039f316a8335bc5f1fa6fc1d: Status 404 returned error can't find the container with id 19539bd948969f6ad4f75c9beef8a27054770fe8039f316a8335bc5f1fa6fc1d Nov 24 07:13:28 crc kubenswrapper[4809]: I1124 07:13:28.466069 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12","Type":"ContainerStarted","Data":"19539bd948969f6ad4f75c9beef8a27054770fe8039f316a8335bc5f1fa6fc1d"} Nov 24 07:13:28 crc kubenswrapper[4809]: I1124 07:13:28.467643 4809 generic.go:334] "Generic (PLEG): container finished" podID="b6217b72-33d3-4b2e-86a2-10bd847003e4" containerID="c02fa30371d4010318a10fd542ad941354969f99e6d49d8655b2cc7d516dc5ed" exitCode=0 Nov 24 07:13:28 crc kubenswrapper[4809]: I1124 07:13:28.467694 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-779kx" event={"ID":"b6217b72-33d3-4b2e-86a2-10bd847003e4","Type":"ContainerDied","Data":"c02fa30371d4010318a10fd542ad941354969f99e6d49d8655b2cc7d516dc5ed"} Nov 24 07:13:28 crc kubenswrapper[4809]: I1124 07:13:28.469607 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"91d79979-625d-4ad3-bbb5-f3f35e7142a4","Type":"ContainerStarted","Data":"1a7fa33c9744e62e6d15b59ec836d0fb6a2d45123006a12a704eafcc6f0d6b97"} Nov 24 07:13:28 crc kubenswrapper[4809]: I1124 07:13:28.909858 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de9cfa7f-1c42-4acc-bca3-d750e80ca2df" path="/var/lib/kubelet/pods/de9cfa7f-1c42-4acc-bca3-d750e80ca2df/volumes" Nov 24 07:13:29 crc kubenswrapper[4809]: I1124 07:13:29.489494 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12","Type":"ContainerStarted","Data":"93e67478c477d5ae12fc6cc99f0c086e265f13bd9e6f24b78121eef62cfe436d"} Nov 24 07:13:29 crc kubenswrapper[4809]: I1124 07:13:29.759602 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-779kx" Nov 24 07:13:29 crc kubenswrapper[4809]: I1124 07:13:29.845904 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6217b72-33d3-4b2e-86a2-10bd847003e4-config-data\") pod \"b6217b72-33d3-4b2e-86a2-10bd847003e4\" (UID: \"b6217b72-33d3-4b2e-86a2-10bd847003e4\") " Nov 24 07:13:29 crc kubenswrapper[4809]: I1124 07:13:29.846134 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6217b72-33d3-4b2e-86a2-10bd847003e4-combined-ca-bundle\") pod \"b6217b72-33d3-4b2e-86a2-10bd847003e4\" (UID: \"b6217b72-33d3-4b2e-86a2-10bd847003e4\") " Nov 24 07:13:29 crc kubenswrapper[4809]: I1124 07:13:29.846185 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqwwr\" (UniqueName: \"kubernetes.io/projected/b6217b72-33d3-4b2e-86a2-10bd847003e4-kube-api-access-bqwwr\") pod \"b6217b72-33d3-4b2e-86a2-10bd847003e4\" (UID: \"b6217b72-33d3-4b2e-86a2-10bd847003e4\") " Nov 24 07:13:29 crc kubenswrapper[4809]: I1124 07:13:29.846218 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6217b72-33d3-4b2e-86a2-10bd847003e4-scripts\") pod \"b6217b72-33d3-4b2e-86a2-10bd847003e4\" (UID: \"b6217b72-33d3-4b2e-86a2-10bd847003e4\") " Nov 24 07:13:29 crc kubenswrapper[4809]: I1124 07:13:29.851055 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6217b72-33d3-4b2e-86a2-10bd847003e4-kube-api-access-bqwwr" (OuterVolumeSpecName: "kube-api-access-bqwwr") pod "b6217b72-33d3-4b2e-86a2-10bd847003e4" (UID: "b6217b72-33d3-4b2e-86a2-10bd847003e4"). InnerVolumeSpecName "kube-api-access-bqwwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:13:29 crc kubenswrapper[4809]: I1124 07:13:29.856734 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6217b72-33d3-4b2e-86a2-10bd847003e4-scripts" (OuterVolumeSpecName: "scripts") pod "b6217b72-33d3-4b2e-86a2-10bd847003e4" (UID: "b6217b72-33d3-4b2e-86a2-10bd847003e4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:29 crc kubenswrapper[4809]: I1124 07:13:29.870822 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6217b72-33d3-4b2e-86a2-10bd847003e4-config-data" (OuterVolumeSpecName: "config-data") pod "b6217b72-33d3-4b2e-86a2-10bd847003e4" (UID: "b6217b72-33d3-4b2e-86a2-10bd847003e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:29 crc kubenswrapper[4809]: I1124 07:13:29.872338 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6217b72-33d3-4b2e-86a2-10bd847003e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b6217b72-33d3-4b2e-86a2-10bd847003e4" (UID: "b6217b72-33d3-4b2e-86a2-10bd847003e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:29 crc kubenswrapper[4809]: I1124 07:13:29.948421 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6217b72-33d3-4b2e-86a2-10bd847003e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:29 crc kubenswrapper[4809]: I1124 07:13:29.948459 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqwwr\" (UniqueName: \"kubernetes.io/projected/b6217b72-33d3-4b2e-86a2-10bd847003e4-kube-api-access-bqwwr\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:29 crc kubenswrapper[4809]: I1124 07:13:29.948477 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6217b72-33d3-4b2e-86a2-10bd847003e4-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:29 crc kubenswrapper[4809]: I1124 07:13:29.948488 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6217b72-33d3-4b2e-86a2-10bd847003e4-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:30 crc kubenswrapper[4809]: I1124 07:13:30.505950 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12","Type":"ContainerStarted","Data":"f1eeac8352d8ca6d2059eaf6fa90f5fb9118557c95a77f09372f6772009b26d6"} Nov 24 07:13:30 crc kubenswrapper[4809]: I1124 07:13:30.508900 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-779kx" Nov 24 07:13:30 crc kubenswrapper[4809]: I1124 07:13:30.519784 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-779kx" event={"ID":"b6217b72-33d3-4b2e-86a2-10bd847003e4","Type":"ContainerDied","Data":"dc12cd3263a05dfe2e57f27f601982352fd2c8c13e510b83f9dc3cb326b4d132"} Nov 24 07:13:30 crc kubenswrapper[4809]: I1124 07:13:30.519839 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc12cd3263a05dfe2e57f27f601982352fd2c8c13e510b83f9dc3cb326b4d132" Nov 24 07:13:30 crc kubenswrapper[4809]: I1124 07:13:30.637725 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 07:13:30 crc kubenswrapper[4809]: E1124 07:13:30.638227 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6217b72-33d3-4b2e-86a2-10bd847003e4" containerName="nova-cell0-conductor-db-sync" Nov 24 07:13:30 crc kubenswrapper[4809]: I1124 07:13:30.638247 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6217b72-33d3-4b2e-86a2-10bd847003e4" containerName="nova-cell0-conductor-db-sync" Nov 24 07:13:30 crc kubenswrapper[4809]: I1124 07:13:30.638489 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6217b72-33d3-4b2e-86a2-10bd847003e4" containerName="nova-cell0-conductor-db-sync" Nov 24 07:13:30 crc kubenswrapper[4809]: I1124 07:13:30.639266 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 07:13:30 crc kubenswrapper[4809]: I1124 07:13:30.640743 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-kmj5v" Nov 24 07:13:30 crc kubenswrapper[4809]: I1124 07:13:30.641543 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 07:13:30 crc kubenswrapper[4809]: I1124 07:13:30.648870 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 07:13:30 crc kubenswrapper[4809]: I1124 07:13:30.764081 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/663c08e1-fd97-42fd-b882-557ca9d71cdf-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"663c08e1-fd97-42fd-b882-557ca9d71cdf\") " pod="openstack/nova-cell0-conductor-0" Nov 24 07:13:30 crc kubenswrapper[4809]: I1124 07:13:30.764409 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzwzv\" (UniqueName: \"kubernetes.io/projected/663c08e1-fd97-42fd-b882-557ca9d71cdf-kube-api-access-kzwzv\") pod \"nova-cell0-conductor-0\" (UID: \"663c08e1-fd97-42fd-b882-557ca9d71cdf\") " pod="openstack/nova-cell0-conductor-0" Nov 24 07:13:30 crc kubenswrapper[4809]: I1124 07:13:30.764677 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/663c08e1-fd97-42fd-b882-557ca9d71cdf-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"663c08e1-fd97-42fd-b882-557ca9d71cdf\") " pod="openstack/nova-cell0-conductor-0" Nov 24 07:13:30 crc kubenswrapper[4809]: I1124 07:13:30.866735 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/663c08e1-fd97-42fd-b882-557ca9d71cdf-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"663c08e1-fd97-42fd-b882-557ca9d71cdf\") " pod="openstack/nova-cell0-conductor-0" Nov 24 07:13:30 crc kubenswrapper[4809]: I1124 07:13:30.866834 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzwzv\" (UniqueName: \"kubernetes.io/projected/663c08e1-fd97-42fd-b882-557ca9d71cdf-kube-api-access-kzwzv\") pod \"nova-cell0-conductor-0\" (UID: \"663c08e1-fd97-42fd-b882-557ca9d71cdf\") " pod="openstack/nova-cell0-conductor-0" Nov 24 07:13:30 crc kubenswrapper[4809]: I1124 07:13:30.866893 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/663c08e1-fd97-42fd-b882-557ca9d71cdf-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"663c08e1-fd97-42fd-b882-557ca9d71cdf\") " pod="openstack/nova-cell0-conductor-0" Nov 24 07:13:30 crc kubenswrapper[4809]: I1124 07:13:30.871145 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/663c08e1-fd97-42fd-b882-557ca9d71cdf-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"663c08e1-fd97-42fd-b882-557ca9d71cdf\") " pod="openstack/nova-cell0-conductor-0" Nov 24 07:13:30 crc kubenswrapper[4809]: I1124 07:13:30.871832 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/663c08e1-fd97-42fd-b882-557ca9d71cdf-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"663c08e1-fd97-42fd-b882-557ca9d71cdf\") " pod="openstack/nova-cell0-conductor-0" Nov 24 07:13:30 crc kubenswrapper[4809]: I1124 07:13:30.886753 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzwzv\" (UniqueName: \"kubernetes.io/projected/663c08e1-fd97-42fd-b882-557ca9d71cdf-kube-api-access-kzwzv\") pod \"nova-cell0-conductor-0\" (UID: \"663c08e1-fd97-42fd-b882-557ca9d71cdf\") " pod="openstack/nova-cell0-conductor-0" Nov 24 07:13:30 crc kubenswrapper[4809]: I1124 07:13:30.960571 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 07:13:30 crc kubenswrapper[4809]: I1124 07:13:30.976517 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 07:13:30 crc kubenswrapper[4809]: I1124 07:13:30.976564 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 07:13:31 crc kubenswrapper[4809]: I1124 07:13:31.021231 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 07:13:31 crc kubenswrapper[4809]: I1124 07:13:31.027390 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 07:13:31 crc kubenswrapper[4809]: I1124 07:13:31.396032 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 07:13:31 crc kubenswrapper[4809]: W1124 07:13:31.410865 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod663c08e1_fd97_42fd_b882_557ca9d71cdf.slice/crio-ca0cb301eb42e76ab3aa65d1cbb13684aaabdec445ec2ecba3fd1a620be3cfcf WatchSource:0}: Error finding container ca0cb301eb42e76ab3aa65d1cbb13684aaabdec445ec2ecba3fd1a620be3cfcf: Status 404 returned error can't find the container with id ca0cb301eb42e76ab3aa65d1cbb13684aaabdec445ec2ecba3fd1a620be3cfcf Nov 24 07:13:31 crc kubenswrapper[4809]: I1124 07:13:31.519842 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12","Type":"ContainerStarted","Data":"c6067d0e403b412b6a7cc202c5711bae48387b5dc487a7ffd0329dcff4961e1c"} Nov 24 07:13:31 crc kubenswrapper[4809]: I1124 07:13:31.521483 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"663c08e1-fd97-42fd-b882-557ca9d71cdf","Type":"ContainerStarted","Data":"ca0cb301eb42e76ab3aa65d1cbb13684aaabdec445ec2ecba3fd1a620be3cfcf"} Nov 24 07:13:31 crc kubenswrapper[4809]: I1124 07:13:31.522824 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 07:13:31 crc kubenswrapper[4809]: I1124 07:13:31.522882 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 07:13:32 crc kubenswrapper[4809]: I1124 07:13:32.533245 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12","Type":"ContainerStarted","Data":"b5b040ad8b4bc04437ba64141bdc8a0c315c0b4d8dbd58823f00a8966850e717"} Nov 24 07:13:32 crc kubenswrapper[4809]: I1124 07:13:32.533902 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 07:13:32 crc kubenswrapper[4809]: I1124 07:13:32.536216 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"663c08e1-fd97-42fd-b882-557ca9d71cdf","Type":"ContainerStarted","Data":"1d2b0560ef47c5bed313f0d1a33e59e8e93ff103e70efb7e4eda440d0b20887c"} Nov 24 07:13:32 crc kubenswrapper[4809]: I1124 07:13:32.536284 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 24 07:13:32 crc kubenswrapper[4809]: I1124 07:13:32.560092 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.861544343 podStartE2EDuration="5.558201728s" podCreationTimestamp="2025-11-24 07:13:27 +0000 UTC" firstStartedPulling="2025-11-24 07:13:28.333220596 +0000 UTC m=+1108.233812301" lastFinishedPulling="2025-11-24 07:13:32.029877971 +0000 UTC m=+1111.930469686" observedRunningTime="2025-11-24 07:13:32.550413091 +0000 UTC m=+1112.451004796" watchObservedRunningTime="2025-11-24 07:13:32.558201728 +0000 UTC m=+1112.458793443" Nov 24 07:13:32 crc kubenswrapper[4809]: I1124 07:13:32.577174 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.577149611 podStartE2EDuration="2.577149611s" podCreationTimestamp="2025-11-24 07:13:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:13:32.569396075 +0000 UTC m=+1112.469987780" watchObservedRunningTime="2025-11-24 07:13:32.577149611 +0000 UTC m=+1112.477741316" Nov 24 07:13:33 crc kubenswrapper[4809]: I1124 07:13:33.299610 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 07:13:33 crc kubenswrapper[4809]: I1124 07:13:33.425576 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 07:13:35 crc kubenswrapper[4809]: I1124 07:13:35.804491 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 07:13:35 crc kubenswrapper[4809]: I1124 07:13:35.804852 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 07:13:35 crc kubenswrapper[4809]: I1124 07:13:35.840238 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 07:13:35 crc kubenswrapper[4809]: I1124 07:13:35.851646 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 07:13:36 crc kubenswrapper[4809]: I1124 07:13:36.572565 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 07:13:36 crc kubenswrapper[4809]: I1124 07:13:36.573385 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 07:13:38 crc kubenswrapper[4809]: I1124 07:13:38.399855 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 07:13:38 crc kubenswrapper[4809]: I1124 07:13:38.410069 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 07:13:40 crc kubenswrapper[4809]: I1124 07:13:40.999058 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.451847 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-8wwqj"] Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.453612 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-8wwqj" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.455859 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.456004 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.506017 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-8wwqj"] Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.595141 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbk6j\" (UniqueName: \"kubernetes.io/projected/825d5761-5b87-4e12-8e7c-392e252331e0-kube-api-access-bbk6j\") pod \"nova-cell0-cell-mapping-8wwqj\" (UID: \"825d5761-5b87-4e12-8e7c-392e252331e0\") " pod="openstack/nova-cell0-cell-mapping-8wwqj" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.595183 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/825d5761-5b87-4e12-8e7c-392e252331e0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-8wwqj\" (UID: \"825d5761-5b87-4e12-8e7c-392e252331e0\") " pod="openstack/nova-cell0-cell-mapping-8wwqj" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.595248 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/825d5761-5b87-4e12-8e7c-392e252331e0-scripts\") pod \"nova-cell0-cell-mapping-8wwqj\" (UID: \"825d5761-5b87-4e12-8e7c-392e252331e0\") " pod="openstack/nova-cell0-cell-mapping-8wwqj" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.595315 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/825d5761-5b87-4e12-8e7c-392e252331e0-config-data\") pod \"nova-cell0-cell-mapping-8wwqj\" (UID: \"825d5761-5b87-4e12-8e7c-392e252331e0\") " pod="openstack/nova-cell0-cell-mapping-8wwqj" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.614511 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.615795 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.620414 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.662778 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.697210 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/825d5761-5b87-4e12-8e7c-392e252331e0-config-data\") pod \"nova-cell0-cell-mapping-8wwqj\" (UID: \"825d5761-5b87-4e12-8e7c-392e252331e0\") " pod="openstack/nova-cell0-cell-mapping-8wwqj" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.697319 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbk6j\" (UniqueName: \"kubernetes.io/projected/825d5761-5b87-4e12-8e7c-392e252331e0-kube-api-access-bbk6j\") pod \"nova-cell0-cell-mapping-8wwqj\" (UID: \"825d5761-5b87-4e12-8e7c-392e252331e0\") " pod="openstack/nova-cell0-cell-mapping-8wwqj" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.697340 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/825d5761-5b87-4e12-8e7c-392e252331e0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-8wwqj\" (UID: \"825d5761-5b87-4e12-8e7c-392e252331e0\") " pod="openstack/nova-cell0-cell-mapping-8wwqj" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.697401 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/825d5761-5b87-4e12-8e7c-392e252331e0-scripts\") pod \"nova-cell0-cell-mapping-8wwqj\" (UID: \"825d5761-5b87-4e12-8e7c-392e252331e0\") " pod="openstack/nova-cell0-cell-mapping-8wwqj" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.717854 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/825d5761-5b87-4e12-8e7c-392e252331e0-scripts\") pod \"nova-cell0-cell-mapping-8wwqj\" (UID: \"825d5761-5b87-4e12-8e7c-392e252331e0\") " pod="openstack/nova-cell0-cell-mapping-8wwqj" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.721704 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/825d5761-5b87-4e12-8e7c-392e252331e0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-8wwqj\" (UID: \"825d5761-5b87-4e12-8e7c-392e252331e0\") " pod="openstack/nova-cell0-cell-mapping-8wwqj" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.729657 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbk6j\" (UniqueName: \"kubernetes.io/projected/825d5761-5b87-4e12-8e7c-392e252331e0-kube-api-access-bbk6j\") pod \"nova-cell0-cell-mapping-8wwqj\" (UID: \"825d5761-5b87-4e12-8e7c-392e252331e0\") " pod="openstack/nova-cell0-cell-mapping-8wwqj" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.737362 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/825d5761-5b87-4e12-8e7c-392e252331e0-config-data\") pod \"nova-cell0-cell-mapping-8wwqj\" (UID: \"825d5761-5b87-4e12-8e7c-392e252331e0\") " pod="openstack/nova-cell0-cell-mapping-8wwqj" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.783792 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.784612 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-8wwqj" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.785313 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.791704 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.800983 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb60935a-17c8-445c-b4ab-99786f623acb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fb60935a-17c8-445c-b4ab-99786f623acb\") " pod="openstack/nova-scheduler-0" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.801078 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-777kw\" (UniqueName: \"kubernetes.io/projected/fb60935a-17c8-445c-b4ab-99786f623acb-kube-api-access-777kw\") pod \"nova-scheduler-0\" (UID: \"fb60935a-17c8-445c-b4ab-99786f623acb\") " pod="openstack/nova-scheduler-0" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.801114 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb60935a-17c8-445c-b4ab-99786f623acb-config-data\") pod \"nova-scheduler-0\" (UID: \"fb60935a-17c8-445c-b4ab-99786f623acb\") " pod="openstack/nova-scheduler-0" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.829179 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.892761 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.893872 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.897282 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.902883 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-777kw\" (UniqueName: \"kubernetes.io/projected/fb60935a-17c8-445c-b4ab-99786f623acb-kube-api-access-777kw\") pod \"nova-scheduler-0\" (UID: \"fb60935a-17c8-445c-b4ab-99786f623acb\") " pod="openstack/nova-scheduler-0" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.902953 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb60935a-17c8-445c-b4ab-99786f623acb-config-data\") pod \"nova-scheduler-0\" (UID: \"fb60935a-17c8-445c-b4ab-99786f623acb\") " pod="openstack/nova-scheduler-0" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.903040 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxfgf\" (UniqueName: \"kubernetes.io/projected/74e55454-2ca0-4059-b284-8b2bee25d853-kube-api-access-cxfgf\") pod \"nova-metadata-0\" (UID: \"74e55454-2ca0-4059-b284-8b2bee25d853\") " pod="openstack/nova-metadata-0" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.903077 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74e55454-2ca0-4059-b284-8b2bee25d853-logs\") pod \"nova-metadata-0\" (UID: \"74e55454-2ca0-4059-b284-8b2bee25d853\") " pod="openstack/nova-metadata-0" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.903121 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74e55454-2ca0-4059-b284-8b2bee25d853-config-data\") pod \"nova-metadata-0\" (UID: \"74e55454-2ca0-4059-b284-8b2bee25d853\") " pod="openstack/nova-metadata-0" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.903164 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74e55454-2ca0-4059-b284-8b2bee25d853-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"74e55454-2ca0-4059-b284-8b2bee25d853\") " pod="openstack/nova-metadata-0" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.905082 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb60935a-17c8-445c-b4ab-99786f623acb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fb60935a-17c8-445c-b4ab-99786f623acb\") " pod="openstack/nova-scheduler-0" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.908043 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb60935a-17c8-445c-b4ab-99786f623acb-config-data\") pod \"nova-scheduler-0\" (UID: \"fb60935a-17c8-445c-b4ab-99786f623acb\") " pod="openstack/nova-scheduler-0" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.909205 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb60935a-17c8-445c-b4ab-99786f623acb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fb60935a-17c8-445c-b4ab-99786f623acb\") " pod="openstack/nova-scheduler-0" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.926786 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.927881 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-777kw\" (UniqueName: \"kubernetes.io/projected/fb60935a-17c8-445c-b4ab-99786f623acb-kube-api-access-777kw\") pod \"nova-scheduler-0\" (UID: \"fb60935a-17c8-445c-b4ab-99786f623acb\") " pod="openstack/nova-scheduler-0" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.928253 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.929939 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.960422 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-jw4ks"] Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.965568 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.974114 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.982248 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 07:13:41 crc kubenswrapper[4809]: I1124 07:13:41.990062 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.001006 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-jw4ks"] Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.006608 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74e55454-2ca0-4059-b284-8b2bee25d853-logs\") pod \"nova-metadata-0\" (UID: \"74e55454-2ca0-4059-b284-8b2bee25d853\") " pod="openstack/nova-metadata-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.006657 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/583f5bd5-82d6-4d48-819a-3f84049011c9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"583f5bd5-82d6-4d48-819a-3f84049011c9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.006699 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74e55454-2ca0-4059-b284-8b2bee25d853-config-data\") pod \"nova-metadata-0\" (UID: \"74e55454-2ca0-4059-b284-8b2bee25d853\") " pod="openstack/nova-metadata-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.006740 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74e55454-2ca0-4059-b284-8b2bee25d853-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"74e55454-2ca0-4059-b284-8b2bee25d853\") " pod="openstack/nova-metadata-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.007669 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74e55454-2ca0-4059-b284-8b2bee25d853-logs\") pod \"nova-metadata-0\" (UID: \"74e55454-2ca0-4059-b284-8b2bee25d853\") " pod="openstack/nova-metadata-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.028769 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/583f5bd5-82d6-4d48-819a-3f84049011c9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"583f5bd5-82d6-4d48-819a-3f84049011c9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.028883 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5b2wn\" (UniqueName: \"kubernetes.io/projected/583f5bd5-82d6-4d48-819a-3f84049011c9-kube-api-access-5b2wn\") pod \"nova-cell1-novncproxy-0\" (UID: \"583f5bd5-82d6-4d48-819a-3f84049011c9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.029200 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxfgf\" (UniqueName: \"kubernetes.io/projected/74e55454-2ca0-4059-b284-8b2bee25d853-kube-api-access-cxfgf\") pod \"nova-metadata-0\" (UID: \"74e55454-2ca0-4059-b284-8b2bee25d853\") " pod="openstack/nova-metadata-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.040682 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74e55454-2ca0-4059-b284-8b2bee25d853-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"74e55454-2ca0-4059-b284-8b2bee25d853\") " pod="openstack/nova-metadata-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.045840 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxfgf\" (UniqueName: \"kubernetes.io/projected/74e55454-2ca0-4059-b284-8b2bee25d853-kube-api-access-cxfgf\") pod \"nova-metadata-0\" (UID: \"74e55454-2ca0-4059-b284-8b2bee25d853\") " pod="openstack/nova-metadata-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.049224 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74e55454-2ca0-4059-b284-8b2bee25d853-config-data\") pod \"nova-metadata-0\" (UID: \"74e55454-2ca0-4059-b284-8b2bee25d853\") " pod="openstack/nova-metadata-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.132841 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bxqb\" (UniqueName: \"kubernetes.io/projected/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-kube-api-access-9bxqb\") pod \"nova-api-0\" (UID: \"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c\") " pod="openstack/nova-api-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.132901 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-jw4ks\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.132942 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5b2wn\" (UniqueName: \"kubernetes.io/projected/583f5bd5-82d6-4d48-819a-3f84049011c9-kube-api-access-5b2wn\") pod \"nova-cell1-novncproxy-0\" (UID: \"583f5bd5-82d6-4d48-819a-3f84049011c9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.132981 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-config-data\") pod \"nova-api-0\" (UID: \"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c\") " pod="openstack/nova-api-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.133017 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-config\") pod \"dnsmasq-dns-757b4f8459-jw4ks\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.133047 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/583f5bd5-82d6-4d48-819a-3f84049011c9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"583f5bd5-82d6-4d48-819a-3f84049011c9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.133077 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c\") " pod="openstack/nova-api-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.133102 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-logs\") pod \"nova-api-0\" (UID: \"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c\") " pod="openstack/nova-api-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.133132 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbjsp\" (UniqueName: \"kubernetes.io/projected/9feafcbd-4cea-47dc-b276-6878ad28e569-kube-api-access-fbjsp\") pod \"dnsmasq-dns-757b4f8459-jw4ks\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.133173 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-dns-svc\") pod \"dnsmasq-dns-757b4f8459-jw4ks\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.133203 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-jw4ks\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.133228 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/583f5bd5-82d6-4d48-819a-3f84049011c9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"583f5bd5-82d6-4d48-819a-3f84049011c9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.133250 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-jw4ks\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.138144 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/583f5bd5-82d6-4d48-819a-3f84049011c9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"583f5bd5-82d6-4d48-819a-3f84049011c9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.139572 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/583f5bd5-82d6-4d48-819a-3f84049011c9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"583f5bd5-82d6-4d48-819a-3f84049011c9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.150108 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5b2wn\" (UniqueName: \"kubernetes.io/projected/583f5bd5-82d6-4d48-819a-3f84049011c9-kube-api-access-5b2wn\") pod \"nova-cell1-novncproxy-0\" (UID: \"583f5bd5-82d6-4d48-819a-3f84049011c9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.195492 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.234125 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-jw4ks\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.234180 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-config-data\") pod \"nova-api-0\" (UID: \"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c\") " pod="openstack/nova-api-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.234209 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-config\") pod \"dnsmasq-dns-757b4f8459-jw4ks\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.234241 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c\") " pod="openstack/nova-api-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.234264 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-logs\") pod \"nova-api-0\" (UID: \"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c\") " pod="openstack/nova-api-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.234288 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbjsp\" (UniqueName: \"kubernetes.io/projected/9feafcbd-4cea-47dc-b276-6878ad28e569-kube-api-access-fbjsp\") pod \"dnsmasq-dns-757b4f8459-jw4ks\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.234322 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-dns-svc\") pod \"dnsmasq-dns-757b4f8459-jw4ks\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.234347 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-jw4ks\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.234413 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-jw4ks\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.234441 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bxqb\" (UniqueName: \"kubernetes.io/projected/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-kube-api-access-9bxqb\") pod \"nova-api-0\" (UID: \"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c\") " pod="openstack/nova-api-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.235586 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-logs\") pod \"nova-api-0\" (UID: \"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c\") " pod="openstack/nova-api-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.235749 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-jw4ks\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.236406 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-dns-svc\") pod \"dnsmasq-dns-757b4f8459-jw4ks\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.236913 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-jw4ks\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.237403 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-jw4ks\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.238436 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-config\") pod \"dnsmasq-dns-757b4f8459-jw4ks\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.240879 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-config-data\") pod \"nova-api-0\" (UID: \"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c\") " pod="openstack/nova-api-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.244582 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c\") " pod="openstack/nova-api-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.253316 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbjsp\" (UniqueName: \"kubernetes.io/projected/9feafcbd-4cea-47dc-b276-6878ad28e569-kube-api-access-fbjsp\") pod \"dnsmasq-dns-757b4f8459-jw4ks\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.260624 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bxqb\" (UniqueName: \"kubernetes.io/projected/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-kube-api-access-9bxqb\") pod \"nova-api-0\" (UID: \"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c\") " pod="openstack/nova-api-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.267512 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.276545 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.308131 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.390567 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-8wwqj"] Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.540016 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.637097 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9fqbw"] Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.638473 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9fqbw" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.641764 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.641845 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.643168 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsxwr\" (UniqueName: \"kubernetes.io/projected/45003542-95f6-46b1-8ff9-06dfc99c1d93-kube-api-access-hsxwr\") pod \"nova-cell1-conductor-db-sync-9fqbw\" (UID: \"45003542-95f6-46b1-8ff9-06dfc99c1d93\") " pod="openstack/nova-cell1-conductor-db-sync-9fqbw" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.643258 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45003542-95f6-46b1-8ff9-06dfc99c1d93-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-9fqbw\" (UID: \"45003542-95f6-46b1-8ff9-06dfc99c1d93\") " pod="openstack/nova-cell1-conductor-db-sync-9fqbw" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.643376 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45003542-95f6-46b1-8ff9-06dfc99c1d93-scripts\") pod \"nova-cell1-conductor-db-sync-9fqbw\" (UID: \"45003542-95f6-46b1-8ff9-06dfc99c1d93\") " pod="openstack/nova-cell1-conductor-db-sync-9fqbw" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.643619 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45003542-95f6-46b1-8ff9-06dfc99c1d93-config-data\") pod \"nova-cell1-conductor-db-sync-9fqbw\" (UID: \"45003542-95f6-46b1-8ff9-06dfc99c1d93\") " pod="openstack/nova-cell1-conductor-db-sync-9fqbw" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.649371 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fb60935a-17c8-445c-b4ab-99786f623acb","Type":"ContainerStarted","Data":"7f04a43a86e56e5ae3244d709a800ca90776dda5a055ef4469923707ab5cf68f"} Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.650912 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-8wwqj" event={"ID":"825d5761-5b87-4e12-8e7c-392e252331e0","Type":"ContainerStarted","Data":"57004dc8b7f9df4ec8db75cf84969e39d96583e144d57dc2f273b070fed5527f"} Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.651575 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9fqbw"] Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.745906 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45003542-95f6-46b1-8ff9-06dfc99c1d93-config-data\") pod \"nova-cell1-conductor-db-sync-9fqbw\" (UID: \"45003542-95f6-46b1-8ff9-06dfc99c1d93\") " pod="openstack/nova-cell1-conductor-db-sync-9fqbw" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.746118 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsxwr\" (UniqueName: \"kubernetes.io/projected/45003542-95f6-46b1-8ff9-06dfc99c1d93-kube-api-access-hsxwr\") pod \"nova-cell1-conductor-db-sync-9fqbw\" (UID: \"45003542-95f6-46b1-8ff9-06dfc99c1d93\") " pod="openstack/nova-cell1-conductor-db-sync-9fqbw" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.746441 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45003542-95f6-46b1-8ff9-06dfc99c1d93-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-9fqbw\" (UID: \"45003542-95f6-46b1-8ff9-06dfc99c1d93\") " pod="openstack/nova-cell1-conductor-db-sync-9fqbw" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.746875 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45003542-95f6-46b1-8ff9-06dfc99c1d93-scripts\") pod \"nova-cell1-conductor-db-sync-9fqbw\" (UID: \"45003542-95f6-46b1-8ff9-06dfc99c1d93\") " pod="openstack/nova-cell1-conductor-db-sync-9fqbw" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.751611 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45003542-95f6-46b1-8ff9-06dfc99c1d93-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-9fqbw\" (UID: \"45003542-95f6-46b1-8ff9-06dfc99c1d93\") " pod="openstack/nova-cell1-conductor-db-sync-9fqbw" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.751656 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45003542-95f6-46b1-8ff9-06dfc99c1d93-config-data\") pod \"nova-cell1-conductor-db-sync-9fqbw\" (UID: \"45003542-95f6-46b1-8ff9-06dfc99c1d93\") " pod="openstack/nova-cell1-conductor-db-sync-9fqbw" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.754485 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45003542-95f6-46b1-8ff9-06dfc99c1d93-scripts\") pod \"nova-cell1-conductor-db-sync-9fqbw\" (UID: \"45003542-95f6-46b1-8ff9-06dfc99c1d93\") " pod="openstack/nova-cell1-conductor-db-sync-9fqbw" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.788509 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsxwr\" (UniqueName: \"kubernetes.io/projected/45003542-95f6-46b1-8ff9-06dfc99c1d93-kube-api-access-hsxwr\") pod \"nova-cell1-conductor-db-sync-9fqbw\" (UID: \"45003542-95f6-46b1-8ff9-06dfc99c1d93\") " pod="openstack/nova-cell1-conductor-db-sync-9fqbw" Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.827192 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.906341 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.909873 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:13:42 crc kubenswrapper[4809]: I1124 07:13:42.957275 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9fqbw" Nov 24 07:13:43 crc kubenswrapper[4809]: I1124 07:13:43.091975 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-jw4ks"] Nov 24 07:13:43 crc kubenswrapper[4809]: W1124 07:13:43.106878 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9feafcbd_4cea_47dc_b276_6878ad28e569.slice/crio-ecdb8eadfaf9a4e540d1e16b5a977e33073674718337a3d21e2ce610bf76e862 WatchSource:0}: Error finding container ecdb8eadfaf9a4e540d1e16b5a977e33073674718337a3d21e2ce610bf76e862: Status 404 returned error can't find the container with id ecdb8eadfaf9a4e540d1e16b5a977e33073674718337a3d21e2ce610bf76e862 Nov 24 07:13:43 crc kubenswrapper[4809]: I1124 07:13:43.452744 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9fqbw"] Nov 24 07:13:43 crc kubenswrapper[4809]: W1124 07:13:43.459164 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45003542_95f6_46b1_8ff9_06dfc99c1d93.slice/crio-34c81e6a807fa92cda14c6ec661a3150cdfc4d96fbcd41246a4d82955a961820 WatchSource:0}: Error finding container 34c81e6a807fa92cda14c6ec661a3150cdfc4d96fbcd41246a4d82955a961820: Status 404 returned error can't find the container with id 34c81e6a807fa92cda14c6ec661a3150cdfc4d96fbcd41246a4d82955a961820 Nov 24 07:13:43 crc kubenswrapper[4809]: I1124 07:13:43.665091 4809 generic.go:334] "Generic (PLEG): container finished" podID="9feafcbd-4cea-47dc-b276-6878ad28e569" containerID="cffd6f156e3e5b8f3adb7309fb68d474c847d84da55d7163825a8bd94aacc6b7" exitCode=0 Nov 24 07:13:43 crc kubenswrapper[4809]: I1124 07:13:43.665182 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" event={"ID":"9feafcbd-4cea-47dc-b276-6878ad28e569","Type":"ContainerDied","Data":"cffd6f156e3e5b8f3adb7309fb68d474c847d84da55d7163825a8bd94aacc6b7"} Nov 24 07:13:43 crc kubenswrapper[4809]: I1124 07:13:43.665215 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" event={"ID":"9feafcbd-4cea-47dc-b276-6878ad28e569","Type":"ContainerStarted","Data":"ecdb8eadfaf9a4e540d1e16b5a977e33073674718337a3d21e2ce610bf76e862"} Nov 24 07:13:43 crc kubenswrapper[4809]: I1124 07:13:43.668281 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"74e55454-2ca0-4059-b284-8b2bee25d853","Type":"ContainerStarted","Data":"863d9b8a20ddc9eb31646d63b296e74d08da51d2728e1687b83f833afaeb7201"} Nov 24 07:13:43 crc kubenswrapper[4809]: I1124 07:13:43.669710 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-8wwqj" event={"ID":"825d5761-5b87-4e12-8e7c-392e252331e0","Type":"ContainerStarted","Data":"94fd2308eff927bad656d650cf3af2eb5d33358e78b1e4aaeac0735a4459c80c"} Nov 24 07:13:43 crc kubenswrapper[4809]: I1124 07:13:43.670536 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"583f5bd5-82d6-4d48-819a-3f84049011c9","Type":"ContainerStarted","Data":"0efeda542d0bd46d9b4533002aa182edd248c7b9b856b9e738ef8d01c26d8405"} Nov 24 07:13:43 crc kubenswrapper[4809]: I1124 07:13:43.671654 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c","Type":"ContainerStarted","Data":"bf7389cb17017f91f0e763fb12ce005195c09ccd9aba4abcfadae346ab3a4a04"} Nov 24 07:13:43 crc kubenswrapper[4809]: I1124 07:13:43.681764 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9fqbw" event={"ID":"45003542-95f6-46b1-8ff9-06dfc99c1d93","Type":"ContainerStarted","Data":"34c81e6a807fa92cda14c6ec661a3150cdfc4d96fbcd41246a4d82955a961820"} Nov 24 07:13:43 crc kubenswrapper[4809]: I1124 07:13:43.701941 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-8wwqj" podStartSLOduration=2.701750145 podStartE2EDuration="2.701750145s" podCreationTimestamp="2025-11-24 07:13:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:13:43.697155423 +0000 UTC m=+1123.597747148" watchObservedRunningTime="2025-11-24 07:13:43.701750145 +0000 UTC m=+1123.602341850" Nov 24 07:13:44 crc kubenswrapper[4809]: I1124 07:13:44.694178 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9fqbw" event={"ID":"45003542-95f6-46b1-8ff9-06dfc99c1d93","Type":"ContainerStarted","Data":"af048ce9879c779790b2123c8a7427bdd8f5606ae5dac57293e6c491a0d84d8e"} Nov 24 07:13:44 crc kubenswrapper[4809]: I1124 07:13:44.725236 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-9fqbw" podStartSLOduration=2.7252100390000003 podStartE2EDuration="2.725210039s" podCreationTimestamp="2025-11-24 07:13:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:13:44.71734128 +0000 UTC m=+1124.617932985" watchObservedRunningTime="2025-11-24 07:13:44.725210039 +0000 UTC m=+1124.625801754" Nov 24 07:13:44 crc kubenswrapper[4809]: I1124 07:13:44.971929 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 07:13:44 crc kubenswrapper[4809]: I1124 07:13:44.979853 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:13:46 crc kubenswrapper[4809]: I1124 07:13:46.713354 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fb60935a-17c8-445c-b4ab-99786f623acb","Type":"ContainerStarted","Data":"5dc457b232c378d3ca32c708e2c0773ddd5918ba4bdd32c6f2ada08f52dc3b84"} Nov 24 07:13:46 crc kubenswrapper[4809]: I1124 07:13:46.715244 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"74e55454-2ca0-4059-b284-8b2bee25d853","Type":"ContainerStarted","Data":"3f3e46e342a6ae4cc2bff177f7b9cd886cdb6e67cb2b02993ed65ee0aac40d1b"} Nov 24 07:13:46 crc kubenswrapper[4809]: I1124 07:13:46.715301 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"74e55454-2ca0-4059-b284-8b2bee25d853","Type":"ContainerStarted","Data":"f53d9fbbd7375f26ebca6da1c5c48b32bbd13684f75065f89d0ce4d95452dea7"} Nov 24 07:13:46 crc kubenswrapper[4809]: I1124 07:13:46.715332 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="74e55454-2ca0-4059-b284-8b2bee25d853" containerName="nova-metadata-log" containerID="cri-o://f53d9fbbd7375f26ebca6da1c5c48b32bbd13684f75065f89d0ce4d95452dea7" gracePeriod=30 Nov 24 07:13:46 crc kubenswrapper[4809]: I1124 07:13:46.715379 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="74e55454-2ca0-4059-b284-8b2bee25d853" containerName="nova-metadata-metadata" containerID="cri-o://3f3e46e342a6ae4cc2bff177f7b9cd886cdb6e67cb2b02993ed65ee0aac40d1b" gracePeriod=30 Nov 24 07:13:46 crc kubenswrapper[4809]: I1124 07:13:46.719582 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"583f5bd5-82d6-4d48-819a-3f84049011c9","Type":"ContainerStarted","Data":"4cd3d9f60485af0ee52d3e65e02fda14955ef8130cccd9b4978f7329249cf359"} Nov 24 07:13:46 crc kubenswrapper[4809]: I1124 07:13:46.719750 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="583f5bd5-82d6-4d48-819a-3f84049011c9" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://4cd3d9f60485af0ee52d3e65e02fda14955ef8130cccd9b4978f7329249cf359" gracePeriod=30 Nov 24 07:13:46 crc kubenswrapper[4809]: I1124 07:13:46.730279 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c","Type":"ContainerStarted","Data":"af6567b44f00bee9c3dd572b507523c134d390af573d5d4989dd6ce9c78b9065"} Nov 24 07:13:46 crc kubenswrapper[4809]: I1124 07:13:46.730318 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c","Type":"ContainerStarted","Data":"09111d790ecf2eed86fbfcc0c474c956ffdc7d6a7682d190d63d8ba5cc12e8d6"} Nov 24 07:13:46 crc kubenswrapper[4809]: I1124 07:13:46.734014 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" event={"ID":"9feafcbd-4cea-47dc-b276-6878ad28e569","Type":"ContainerStarted","Data":"947d9b54a8eba14ac0782c84ff11b1e09808c99629b722515f0308ae29545aff"} Nov 24 07:13:46 crc kubenswrapper[4809]: I1124 07:13:46.734411 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:13:46 crc kubenswrapper[4809]: I1124 07:13:46.737112 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.038029919 podStartE2EDuration="5.737096137s" podCreationTimestamp="2025-11-24 07:13:41 +0000 UTC" firstStartedPulling="2025-11-24 07:13:42.560112028 +0000 UTC m=+1122.460703733" lastFinishedPulling="2025-11-24 07:13:45.259178236 +0000 UTC m=+1125.159769951" observedRunningTime="2025-11-24 07:13:46.729801205 +0000 UTC m=+1126.630392900" watchObservedRunningTime="2025-11-24 07:13:46.737096137 +0000 UTC m=+1126.637687842" Nov 24 07:13:46 crc kubenswrapper[4809]: I1124 07:13:46.764332 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.410020643 podStartE2EDuration="5.764309219s" podCreationTimestamp="2025-11-24 07:13:41 +0000 UTC" firstStartedPulling="2025-11-24 07:13:42.906241795 +0000 UTC m=+1122.806833500" lastFinishedPulling="2025-11-24 07:13:45.260530351 +0000 UTC m=+1125.161122076" observedRunningTime="2025-11-24 07:13:46.74887459 +0000 UTC m=+1126.649466305" watchObservedRunningTime="2025-11-24 07:13:46.764309219 +0000 UTC m=+1126.664900924" Nov 24 07:13:46 crc kubenswrapper[4809]: I1124 07:13:46.783927 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.336111363 podStartE2EDuration="5.783869878s" podCreationTimestamp="2025-11-24 07:13:41 +0000 UTC" firstStartedPulling="2025-11-24 07:13:42.82082369 +0000 UTC m=+1122.721415395" lastFinishedPulling="2025-11-24 07:13:45.268582205 +0000 UTC m=+1125.169173910" observedRunningTime="2025-11-24 07:13:46.772582539 +0000 UTC m=+1126.673174254" watchObservedRunningTime="2025-11-24 07:13:46.783869878 +0000 UTC m=+1126.684461623" Nov 24 07:13:46 crc kubenswrapper[4809]: I1124 07:13:46.800957 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.468058931 podStartE2EDuration="5.80093875s" podCreationTimestamp="2025-11-24 07:13:41 +0000 UTC" firstStartedPulling="2025-11-24 07:13:42.930422326 +0000 UTC m=+1122.831014031" lastFinishedPulling="2025-11-24 07:13:45.263302135 +0000 UTC m=+1125.163893850" observedRunningTime="2025-11-24 07:13:46.788374398 +0000 UTC m=+1126.688966113" watchObservedRunningTime="2025-11-24 07:13:46.80093875 +0000 UTC m=+1126.701530465" Nov 24 07:13:46 crc kubenswrapper[4809]: I1124 07:13:46.809759 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" podStartSLOduration=5.809744134 podStartE2EDuration="5.809744134s" podCreationTimestamp="2025-11-24 07:13:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:13:46.806408115 +0000 UTC m=+1126.706999830" watchObservedRunningTime="2025-11-24 07:13:46.809744134 +0000 UTC m=+1126.710335849" Nov 24 07:13:46 crc kubenswrapper[4809]: I1124 07:13:46.966073 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.197060 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.197394 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.239430 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.268056 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.339672 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74e55454-2ca0-4059-b284-8b2bee25d853-config-data\") pod \"74e55454-2ca0-4059-b284-8b2bee25d853\" (UID: \"74e55454-2ca0-4059-b284-8b2bee25d853\") " Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.339799 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxfgf\" (UniqueName: \"kubernetes.io/projected/74e55454-2ca0-4059-b284-8b2bee25d853-kube-api-access-cxfgf\") pod \"74e55454-2ca0-4059-b284-8b2bee25d853\" (UID: \"74e55454-2ca0-4059-b284-8b2bee25d853\") " Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.339915 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74e55454-2ca0-4059-b284-8b2bee25d853-combined-ca-bundle\") pod \"74e55454-2ca0-4059-b284-8b2bee25d853\" (UID: \"74e55454-2ca0-4059-b284-8b2bee25d853\") " Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.339939 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74e55454-2ca0-4059-b284-8b2bee25d853-logs\") pod \"74e55454-2ca0-4059-b284-8b2bee25d853\" (UID: \"74e55454-2ca0-4059-b284-8b2bee25d853\") " Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.341741 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74e55454-2ca0-4059-b284-8b2bee25d853-logs" (OuterVolumeSpecName: "logs") pod "74e55454-2ca0-4059-b284-8b2bee25d853" (UID: "74e55454-2ca0-4059-b284-8b2bee25d853"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.348428 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74e55454-2ca0-4059-b284-8b2bee25d853-kube-api-access-cxfgf" (OuterVolumeSpecName: "kube-api-access-cxfgf") pod "74e55454-2ca0-4059-b284-8b2bee25d853" (UID: "74e55454-2ca0-4059-b284-8b2bee25d853"). InnerVolumeSpecName "kube-api-access-cxfgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.378495 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74e55454-2ca0-4059-b284-8b2bee25d853-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74e55454-2ca0-4059-b284-8b2bee25d853" (UID: "74e55454-2ca0-4059-b284-8b2bee25d853"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.382162 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74e55454-2ca0-4059-b284-8b2bee25d853-config-data" (OuterVolumeSpecName: "config-data") pod "74e55454-2ca0-4059-b284-8b2bee25d853" (UID: "74e55454-2ca0-4059-b284-8b2bee25d853"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.442683 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxfgf\" (UniqueName: \"kubernetes.io/projected/74e55454-2ca0-4059-b284-8b2bee25d853-kube-api-access-cxfgf\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.442739 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74e55454-2ca0-4059-b284-8b2bee25d853-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.442757 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74e55454-2ca0-4059-b284-8b2bee25d853-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.442776 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74e55454-2ca0-4059-b284-8b2bee25d853-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.745368 4809 generic.go:334] "Generic (PLEG): container finished" podID="74e55454-2ca0-4059-b284-8b2bee25d853" containerID="3f3e46e342a6ae4cc2bff177f7b9cd886cdb6e67cb2b02993ed65ee0aac40d1b" exitCode=0 Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.745409 4809 generic.go:334] "Generic (PLEG): container finished" podID="74e55454-2ca0-4059-b284-8b2bee25d853" containerID="f53d9fbbd7375f26ebca6da1c5c48b32bbd13684f75065f89d0ce4d95452dea7" exitCode=143 Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.745435 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.745497 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"74e55454-2ca0-4059-b284-8b2bee25d853","Type":"ContainerDied","Data":"3f3e46e342a6ae4cc2bff177f7b9cd886cdb6e67cb2b02993ed65ee0aac40d1b"} Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.745552 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"74e55454-2ca0-4059-b284-8b2bee25d853","Type":"ContainerDied","Data":"f53d9fbbd7375f26ebca6da1c5c48b32bbd13684f75065f89d0ce4d95452dea7"} Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.745569 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"74e55454-2ca0-4059-b284-8b2bee25d853","Type":"ContainerDied","Data":"863d9b8a20ddc9eb31646d63b296e74d08da51d2728e1687b83f833afaeb7201"} Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.745597 4809 scope.go:117] "RemoveContainer" containerID="3f3e46e342a6ae4cc2bff177f7b9cd886cdb6e67cb2b02993ed65ee0aac40d1b" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.788459 4809 scope.go:117] "RemoveContainer" containerID="f53d9fbbd7375f26ebca6da1c5c48b32bbd13684f75065f89d0ce4d95452dea7" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.791602 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.817780 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.822927 4809 scope.go:117] "RemoveContainer" containerID="3f3e46e342a6ae4cc2bff177f7b9cd886cdb6e67cb2b02993ed65ee0aac40d1b" Nov 24 07:13:47 crc kubenswrapper[4809]: E1124 07:13:47.823722 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f3e46e342a6ae4cc2bff177f7b9cd886cdb6e67cb2b02993ed65ee0aac40d1b\": container with ID starting with 3f3e46e342a6ae4cc2bff177f7b9cd886cdb6e67cb2b02993ed65ee0aac40d1b not found: ID does not exist" containerID="3f3e46e342a6ae4cc2bff177f7b9cd886cdb6e67cb2b02993ed65ee0aac40d1b" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.823919 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f3e46e342a6ae4cc2bff177f7b9cd886cdb6e67cb2b02993ed65ee0aac40d1b"} err="failed to get container status \"3f3e46e342a6ae4cc2bff177f7b9cd886cdb6e67cb2b02993ed65ee0aac40d1b\": rpc error: code = NotFound desc = could not find container \"3f3e46e342a6ae4cc2bff177f7b9cd886cdb6e67cb2b02993ed65ee0aac40d1b\": container with ID starting with 3f3e46e342a6ae4cc2bff177f7b9cd886cdb6e67cb2b02993ed65ee0aac40d1b not found: ID does not exist" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.824189 4809 scope.go:117] "RemoveContainer" containerID="f53d9fbbd7375f26ebca6da1c5c48b32bbd13684f75065f89d0ce4d95452dea7" Nov 24 07:13:47 crc kubenswrapper[4809]: E1124 07:13:47.824907 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f53d9fbbd7375f26ebca6da1c5c48b32bbd13684f75065f89d0ce4d95452dea7\": container with ID starting with f53d9fbbd7375f26ebca6da1c5c48b32bbd13684f75065f89d0ce4d95452dea7 not found: ID does not exist" containerID="f53d9fbbd7375f26ebca6da1c5c48b32bbd13684f75065f89d0ce4d95452dea7" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.825133 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f53d9fbbd7375f26ebca6da1c5c48b32bbd13684f75065f89d0ce4d95452dea7"} err="failed to get container status \"f53d9fbbd7375f26ebca6da1c5c48b32bbd13684f75065f89d0ce4d95452dea7\": rpc error: code = NotFound desc = could not find container \"f53d9fbbd7375f26ebca6da1c5c48b32bbd13684f75065f89d0ce4d95452dea7\": container with ID starting with f53d9fbbd7375f26ebca6da1c5c48b32bbd13684f75065f89d0ce4d95452dea7 not found: ID does not exist" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.825281 4809 scope.go:117] "RemoveContainer" containerID="3f3e46e342a6ae4cc2bff177f7b9cd886cdb6e67cb2b02993ed65ee0aac40d1b" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.827291 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f3e46e342a6ae4cc2bff177f7b9cd886cdb6e67cb2b02993ed65ee0aac40d1b"} err="failed to get container status \"3f3e46e342a6ae4cc2bff177f7b9cd886cdb6e67cb2b02993ed65ee0aac40d1b\": rpc error: code = NotFound desc = could not find container \"3f3e46e342a6ae4cc2bff177f7b9cd886cdb6e67cb2b02993ed65ee0aac40d1b\": container with ID starting with 3f3e46e342a6ae4cc2bff177f7b9cd886cdb6e67cb2b02993ed65ee0aac40d1b not found: ID does not exist" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.827489 4809 scope.go:117] "RemoveContainer" containerID="f53d9fbbd7375f26ebca6da1c5c48b32bbd13684f75065f89d0ce4d95452dea7" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.828058 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f53d9fbbd7375f26ebca6da1c5c48b32bbd13684f75065f89d0ce4d95452dea7"} err="failed to get container status \"f53d9fbbd7375f26ebca6da1c5c48b32bbd13684f75065f89d0ce4d95452dea7\": rpc error: code = NotFound desc = could not find container \"f53d9fbbd7375f26ebca6da1c5c48b32bbd13684f75065f89d0ce4d95452dea7\": container with ID starting with f53d9fbbd7375f26ebca6da1c5c48b32bbd13684f75065f89d0ce4d95452dea7 not found: ID does not exist" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.849439 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:13:47 crc kubenswrapper[4809]: E1124 07:13:47.850002 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74e55454-2ca0-4059-b284-8b2bee25d853" containerName="nova-metadata-metadata" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.850022 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="74e55454-2ca0-4059-b284-8b2bee25d853" containerName="nova-metadata-metadata" Nov 24 07:13:47 crc kubenswrapper[4809]: E1124 07:13:47.850051 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74e55454-2ca0-4059-b284-8b2bee25d853" containerName="nova-metadata-log" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.850060 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="74e55454-2ca0-4059-b284-8b2bee25d853" containerName="nova-metadata-log" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.850317 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="74e55454-2ca0-4059-b284-8b2bee25d853" containerName="nova-metadata-log" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.850357 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="74e55454-2ca0-4059-b284-8b2bee25d853" containerName="nova-metadata-metadata" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.851678 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.856507 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.856760 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.864399 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.955316 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56d958c1-a749-4f2f-b9a7-d5d526da77c7-config-data\") pod \"nova-metadata-0\" (UID: \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\") " pod="openstack/nova-metadata-0" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.955365 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/56d958c1-a749-4f2f-b9a7-d5d526da77c7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\") " pod="openstack/nova-metadata-0" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.955394 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56d958c1-a749-4f2f-b9a7-d5d526da77c7-logs\") pod \"nova-metadata-0\" (UID: \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\") " pod="openstack/nova-metadata-0" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.955639 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkz2z\" (UniqueName: \"kubernetes.io/projected/56d958c1-a749-4f2f-b9a7-d5d526da77c7-kube-api-access-zkz2z\") pod \"nova-metadata-0\" (UID: \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\") " pod="openstack/nova-metadata-0" Nov 24 07:13:47 crc kubenswrapper[4809]: I1124 07:13:47.955707 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56d958c1-a749-4f2f-b9a7-d5d526da77c7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\") " pod="openstack/nova-metadata-0" Nov 24 07:13:48 crc kubenswrapper[4809]: I1124 07:13:48.043377 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:13:48 crc kubenswrapper[4809]: I1124 07:13:48.043430 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:13:48 crc kubenswrapper[4809]: I1124 07:13:48.043471 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 07:13:48 crc kubenswrapper[4809]: I1124 07:13:48.044136 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"37feb2230d7f506014799017af7463c0d7873535f5cdb2251d0334f2328d10b8"} pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:13:48 crc kubenswrapper[4809]: I1124 07:13:48.044191 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" containerID="cri-o://37feb2230d7f506014799017af7463c0d7873535f5cdb2251d0334f2328d10b8" gracePeriod=600 Nov 24 07:13:48 crc kubenswrapper[4809]: I1124 07:13:48.057632 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkz2z\" (UniqueName: \"kubernetes.io/projected/56d958c1-a749-4f2f-b9a7-d5d526da77c7-kube-api-access-zkz2z\") pod \"nova-metadata-0\" (UID: \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\") " pod="openstack/nova-metadata-0" Nov 24 07:13:48 crc kubenswrapper[4809]: I1124 07:13:48.057681 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56d958c1-a749-4f2f-b9a7-d5d526da77c7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\") " pod="openstack/nova-metadata-0" Nov 24 07:13:48 crc kubenswrapper[4809]: I1124 07:13:48.057741 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56d958c1-a749-4f2f-b9a7-d5d526da77c7-config-data\") pod \"nova-metadata-0\" (UID: \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\") " pod="openstack/nova-metadata-0" Nov 24 07:13:48 crc kubenswrapper[4809]: I1124 07:13:48.057759 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/56d958c1-a749-4f2f-b9a7-d5d526da77c7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\") " pod="openstack/nova-metadata-0" Nov 24 07:13:48 crc kubenswrapper[4809]: I1124 07:13:48.057775 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56d958c1-a749-4f2f-b9a7-d5d526da77c7-logs\") pod \"nova-metadata-0\" (UID: \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\") " pod="openstack/nova-metadata-0" Nov 24 07:13:48 crc kubenswrapper[4809]: I1124 07:13:48.059800 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56d958c1-a749-4f2f-b9a7-d5d526da77c7-logs\") pod \"nova-metadata-0\" (UID: \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\") " pod="openstack/nova-metadata-0" Nov 24 07:13:48 crc kubenswrapper[4809]: I1124 07:13:48.064513 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/56d958c1-a749-4f2f-b9a7-d5d526da77c7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\") " pod="openstack/nova-metadata-0" Nov 24 07:13:48 crc kubenswrapper[4809]: I1124 07:13:48.067156 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56d958c1-a749-4f2f-b9a7-d5d526da77c7-config-data\") pod \"nova-metadata-0\" (UID: \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\") " pod="openstack/nova-metadata-0" Nov 24 07:13:48 crc kubenswrapper[4809]: I1124 07:13:48.067348 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56d958c1-a749-4f2f-b9a7-d5d526da77c7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\") " pod="openstack/nova-metadata-0" Nov 24 07:13:48 crc kubenswrapper[4809]: I1124 07:13:48.093052 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkz2z\" (UniqueName: \"kubernetes.io/projected/56d958c1-a749-4f2f-b9a7-d5d526da77c7-kube-api-access-zkz2z\") pod \"nova-metadata-0\" (UID: \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\") " pod="openstack/nova-metadata-0" Nov 24 07:13:48 crc kubenswrapper[4809]: I1124 07:13:48.183656 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:13:48 crc kubenswrapper[4809]: I1124 07:13:48.661655 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:13:48 crc kubenswrapper[4809]: W1124 07:13:48.665679 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56d958c1_a749_4f2f_b9a7_d5d526da77c7.slice/crio-9bcd4beda3f2dfd786a4d5ef4e1aa85e3b26da7fd8508e7b0ef15f63db470dc1 WatchSource:0}: Error finding container 9bcd4beda3f2dfd786a4d5ef4e1aa85e3b26da7fd8508e7b0ef15f63db470dc1: Status 404 returned error can't find the container with id 9bcd4beda3f2dfd786a4d5ef4e1aa85e3b26da7fd8508e7b0ef15f63db470dc1 Nov 24 07:13:48 crc kubenswrapper[4809]: I1124 07:13:48.764043 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"56d958c1-a749-4f2f-b9a7-d5d526da77c7","Type":"ContainerStarted","Data":"9bcd4beda3f2dfd786a4d5ef4e1aa85e3b26da7fd8508e7b0ef15f63db470dc1"} Nov 24 07:13:48 crc kubenswrapper[4809]: I1124 07:13:48.771537 4809 generic.go:334] "Generic (PLEG): container finished" podID="be8050d5-1aea-453e-a06e-87702b0e856b" containerID="37feb2230d7f506014799017af7463c0d7873535f5cdb2251d0334f2328d10b8" exitCode=0 Nov 24 07:13:48 crc kubenswrapper[4809]: I1124 07:13:48.771615 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerDied","Data":"37feb2230d7f506014799017af7463c0d7873535f5cdb2251d0334f2328d10b8"} Nov 24 07:13:48 crc kubenswrapper[4809]: I1124 07:13:48.771875 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerStarted","Data":"11616e3dc08f43f601be45c4cf1cbd156c734d2c4a8bf401dd3c381f1cf4aefb"} Nov 24 07:13:48 crc kubenswrapper[4809]: I1124 07:13:48.771896 4809 scope.go:117] "RemoveContainer" containerID="5bbd5cb6a3a9e9de8c1c6f89b55f7b7f733fb4a270390a494ec13ae9d9f8fbca" Nov 24 07:13:48 crc kubenswrapper[4809]: I1124 07:13:48.902598 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74e55454-2ca0-4059-b284-8b2bee25d853" path="/var/lib/kubelet/pods/74e55454-2ca0-4059-b284-8b2bee25d853/volumes" Nov 24 07:13:49 crc kubenswrapper[4809]: I1124 07:13:49.783283 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"56d958c1-a749-4f2f-b9a7-d5d526da77c7","Type":"ContainerStarted","Data":"4a90f038dc639efc80174ba98a119140ecc2f2acc75f680554a55f967ac9ac23"} Nov 24 07:13:49 crc kubenswrapper[4809]: I1124 07:13:49.783928 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"56d958c1-a749-4f2f-b9a7-d5d526da77c7","Type":"ContainerStarted","Data":"5e4fa1f28f7adf3189d1f81512b329afc3e9694a9210ce4734632959889f2840"} Nov 24 07:13:49 crc kubenswrapper[4809]: I1124 07:13:49.785706 4809 generic.go:334] "Generic (PLEG): container finished" podID="825d5761-5b87-4e12-8e7c-392e252331e0" containerID="94fd2308eff927bad656d650cf3af2eb5d33358e78b1e4aaeac0735a4459c80c" exitCode=0 Nov 24 07:13:49 crc kubenswrapper[4809]: I1124 07:13:49.785766 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-8wwqj" event={"ID":"825d5761-5b87-4e12-8e7c-392e252331e0","Type":"ContainerDied","Data":"94fd2308eff927bad656d650cf3af2eb5d33358e78b1e4aaeac0735a4459c80c"} Nov 24 07:13:49 crc kubenswrapper[4809]: I1124 07:13:49.807034 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.807010597 podStartE2EDuration="2.807010597s" podCreationTimestamp="2025-11-24 07:13:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:13:49.803076163 +0000 UTC m=+1129.703667878" watchObservedRunningTime="2025-11-24 07:13:49.807010597 +0000 UTC m=+1129.707602322" Nov 24 07:13:51 crc kubenswrapper[4809]: I1124 07:13:51.256890 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-8wwqj" Nov 24 07:13:51 crc kubenswrapper[4809]: I1124 07:13:51.426544 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbk6j\" (UniqueName: \"kubernetes.io/projected/825d5761-5b87-4e12-8e7c-392e252331e0-kube-api-access-bbk6j\") pod \"825d5761-5b87-4e12-8e7c-392e252331e0\" (UID: \"825d5761-5b87-4e12-8e7c-392e252331e0\") " Nov 24 07:13:51 crc kubenswrapper[4809]: I1124 07:13:51.426798 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/825d5761-5b87-4e12-8e7c-392e252331e0-combined-ca-bundle\") pod \"825d5761-5b87-4e12-8e7c-392e252331e0\" (UID: \"825d5761-5b87-4e12-8e7c-392e252331e0\") " Nov 24 07:13:51 crc kubenswrapper[4809]: I1124 07:13:51.426832 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/825d5761-5b87-4e12-8e7c-392e252331e0-scripts\") pod \"825d5761-5b87-4e12-8e7c-392e252331e0\" (UID: \"825d5761-5b87-4e12-8e7c-392e252331e0\") " Nov 24 07:13:51 crc kubenswrapper[4809]: I1124 07:13:51.426932 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/825d5761-5b87-4e12-8e7c-392e252331e0-config-data\") pod \"825d5761-5b87-4e12-8e7c-392e252331e0\" (UID: \"825d5761-5b87-4e12-8e7c-392e252331e0\") " Nov 24 07:13:51 crc kubenswrapper[4809]: I1124 07:13:51.437920 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/825d5761-5b87-4e12-8e7c-392e252331e0-scripts" (OuterVolumeSpecName: "scripts") pod "825d5761-5b87-4e12-8e7c-392e252331e0" (UID: "825d5761-5b87-4e12-8e7c-392e252331e0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:51 crc kubenswrapper[4809]: I1124 07:13:51.457800 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/825d5761-5b87-4e12-8e7c-392e252331e0-kube-api-access-bbk6j" (OuterVolumeSpecName: "kube-api-access-bbk6j") pod "825d5761-5b87-4e12-8e7c-392e252331e0" (UID: "825d5761-5b87-4e12-8e7c-392e252331e0"). InnerVolumeSpecName "kube-api-access-bbk6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:13:51 crc kubenswrapper[4809]: I1124 07:13:51.459212 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/825d5761-5b87-4e12-8e7c-392e252331e0-config-data" (OuterVolumeSpecName: "config-data") pod "825d5761-5b87-4e12-8e7c-392e252331e0" (UID: "825d5761-5b87-4e12-8e7c-392e252331e0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:51 crc kubenswrapper[4809]: I1124 07:13:51.460934 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/825d5761-5b87-4e12-8e7c-392e252331e0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "825d5761-5b87-4e12-8e7c-392e252331e0" (UID: "825d5761-5b87-4e12-8e7c-392e252331e0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:51 crc kubenswrapper[4809]: I1124 07:13:51.530272 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/825d5761-5b87-4e12-8e7c-392e252331e0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:51 crc kubenswrapper[4809]: I1124 07:13:51.530322 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/825d5761-5b87-4e12-8e7c-392e252331e0-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:51 crc kubenswrapper[4809]: I1124 07:13:51.530336 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/825d5761-5b87-4e12-8e7c-392e252331e0-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:51 crc kubenswrapper[4809]: I1124 07:13:51.530350 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbk6j\" (UniqueName: \"kubernetes.io/projected/825d5761-5b87-4e12-8e7c-392e252331e0-kube-api-access-bbk6j\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:51 crc kubenswrapper[4809]: I1124 07:13:51.821589 4809 generic.go:334] "Generic (PLEG): container finished" podID="45003542-95f6-46b1-8ff9-06dfc99c1d93" containerID="af048ce9879c779790b2123c8a7427bdd8f5606ae5dac57293e6c491a0d84d8e" exitCode=0 Nov 24 07:13:51 crc kubenswrapper[4809]: I1124 07:13:51.821690 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9fqbw" event={"ID":"45003542-95f6-46b1-8ff9-06dfc99c1d93","Type":"ContainerDied","Data":"af048ce9879c779790b2123c8a7427bdd8f5606ae5dac57293e6c491a0d84d8e"} Nov 24 07:13:51 crc kubenswrapper[4809]: I1124 07:13:51.826947 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-8wwqj" event={"ID":"825d5761-5b87-4e12-8e7c-392e252331e0","Type":"ContainerDied","Data":"57004dc8b7f9df4ec8db75cf84969e39d96583e144d57dc2f273b070fed5527f"} Nov 24 07:13:51 crc kubenswrapper[4809]: I1124 07:13:51.827011 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57004dc8b7f9df4ec8db75cf84969e39d96583e144d57dc2f273b070fed5527f" Nov 24 07:13:51 crc kubenswrapper[4809]: I1124 07:13:51.827071 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-8wwqj" Nov 24 07:13:51 crc kubenswrapper[4809]: I1124 07:13:51.967157 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.016202 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.017098 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.034078 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.034518 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b71bf545-2e9d-4dd4-bd8a-db1eb36a535c" containerName="nova-api-api" containerID="cri-o://af6567b44f00bee9c3dd572b507523c134d390af573d5d4989dd6ce9c78b9065" gracePeriod=30 Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.034665 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b71bf545-2e9d-4dd4-bd8a-db1eb36a535c" containerName="nova-api-log" containerID="cri-o://09111d790ecf2eed86fbfcc0c474c956ffdc7d6a7682d190d63d8ba5cc12e8d6" gracePeriod=30 Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.050742 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.051068 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="56d958c1-a749-4f2f-b9a7-d5d526da77c7" containerName="nova-metadata-metadata" containerID="cri-o://4a90f038dc639efc80174ba98a119140ecc2f2acc75f680554a55f967ac9ac23" gracePeriod=30 Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.051223 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="56d958c1-a749-4f2f-b9a7-d5d526da77c7" containerName="nova-metadata-log" containerID="cri-o://5e4fa1f28f7adf3189d1f81512b329afc3e9694a9210ce4734632959889f2840" gracePeriod=30 Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.313159 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.387230 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-br547"] Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.389760 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-br547" podUID="ec7ac87d-2d70-4384-aceb-6f2eca6bf94b" containerName="dnsmasq-dns" containerID="cri-o://a1a35d39a40c6f8fb1fc684916016ba71fa8ba999ca9d9e1802d4c3c9343b4c3" gracePeriod=10 Nov 24 07:13:52 crc kubenswrapper[4809]: E1124 07:13:52.545782 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec7ac87d_2d70_4384_aceb_6f2eca6bf94b.slice/crio-conmon-a1a35d39a40c6f8fb1fc684916016ba71fa8ba999ca9d9e1802d4c3c9343b4c3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec7ac87d_2d70_4384_aceb_6f2eca6bf94b.slice/crio-a1a35d39a40c6f8fb1fc684916016ba71fa8ba999ca9d9e1802d4c3c9343b4c3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb71bf545_2e9d_4dd4_bd8a_db1eb36a535c.slice/crio-conmon-09111d790ecf2eed86fbfcc0c474c956ffdc7d6a7682d190d63d8ba5cc12e8d6.scope\": RecentStats: unable to find data in memory cache]" Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.826159 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.855616 4809 generic.go:334] "Generic (PLEG): container finished" podID="b71bf545-2e9d-4dd4-bd8a-db1eb36a535c" containerID="af6567b44f00bee9c3dd572b507523c134d390af573d5d4989dd6ce9c78b9065" exitCode=0 Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.855651 4809 generic.go:334] "Generic (PLEG): container finished" podID="b71bf545-2e9d-4dd4-bd8a-db1eb36a535c" containerID="09111d790ecf2eed86fbfcc0c474c956ffdc7d6a7682d190d63d8ba5cc12e8d6" exitCode=143 Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.855720 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c","Type":"ContainerDied","Data":"af6567b44f00bee9c3dd572b507523c134d390af573d5d4989dd6ce9c78b9065"} Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.855769 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c","Type":"ContainerDied","Data":"09111d790ecf2eed86fbfcc0c474c956ffdc7d6a7682d190d63d8ba5cc12e8d6"} Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.863820 4809 generic.go:334] "Generic (PLEG): container finished" podID="ec7ac87d-2d70-4384-aceb-6f2eca6bf94b" containerID="a1a35d39a40c6f8fb1fc684916016ba71fa8ba999ca9d9e1802d4c3c9343b4c3" exitCode=0 Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.863910 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-br547" Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.864178 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-br547" event={"ID":"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b","Type":"ContainerDied","Data":"a1a35d39a40c6f8fb1fc684916016ba71fa8ba999ca9d9e1802d4c3c9343b4c3"} Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.864210 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-br547" event={"ID":"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b","Type":"ContainerDied","Data":"3ed4b8b95947ba5dffe2a86040a1548d29f83b6a1a0bbda277ef96288140d4d6"} Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.864226 4809 scope.go:117] "RemoveContainer" containerID="a1a35d39a40c6f8fb1fc684916016ba71fa8ba999ca9d9e1802d4c3c9343b4c3" Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.871384 4809 generic.go:334] "Generic (PLEG): container finished" podID="56d958c1-a749-4f2f-b9a7-d5d526da77c7" containerID="4a90f038dc639efc80174ba98a119140ecc2f2acc75f680554a55f967ac9ac23" exitCode=0 Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.871406 4809 generic.go:334] "Generic (PLEG): container finished" podID="56d958c1-a749-4f2f-b9a7-d5d526da77c7" containerID="5e4fa1f28f7adf3189d1f81512b329afc3e9694a9210ce4734632959889f2840" exitCode=143 Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.871569 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"56d958c1-a749-4f2f-b9a7-d5d526da77c7","Type":"ContainerDied","Data":"4a90f038dc639efc80174ba98a119140ecc2f2acc75f680554a55f967ac9ac23"} Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.871594 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"56d958c1-a749-4f2f-b9a7-d5d526da77c7","Type":"ContainerDied","Data":"5e4fa1f28f7adf3189d1f81512b329afc3e9694a9210ce4734632959889f2840"} Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.905083 4809 scope.go:117] "RemoveContainer" containerID="110725378d5569ef98bf845f63c45738e84831d095aa4ac8049621bdef38cd52" Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.923542 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.939095 4809 scope.go:117] "RemoveContainer" containerID="a1a35d39a40c6f8fb1fc684916016ba71fa8ba999ca9d9e1802d4c3c9343b4c3" Nov 24 07:13:52 crc kubenswrapper[4809]: E1124 07:13:52.939693 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1a35d39a40c6f8fb1fc684916016ba71fa8ba999ca9d9e1802d4c3c9343b4c3\": container with ID starting with a1a35d39a40c6f8fb1fc684916016ba71fa8ba999ca9d9e1802d4c3c9343b4c3 not found: ID does not exist" containerID="a1a35d39a40c6f8fb1fc684916016ba71fa8ba999ca9d9e1802d4c3c9343b4c3" Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.939728 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1a35d39a40c6f8fb1fc684916016ba71fa8ba999ca9d9e1802d4c3c9343b4c3"} err="failed to get container status \"a1a35d39a40c6f8fb1fc684916016ba71fa8ba999ca9d9e1802d4c3c9343b4c3\": rpc error: code = NotFound desc = could not find container \"a1a35d39a40c6f8fb1fc684916016ba71fa8ba999ca9d9e1802d4c3c9343b4c3\": container with ID starting with a1a35d39a40c6f8fb1fc684916016ba71fa8ba999ca9d9e1802d4c3c9343b4c3 not found: ID does not exist" Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.939776 4809 scope.go:117] "RemoveContainer" containerID="110725378d5569ef98bf845f63c45738e84831d095aa4ac8049621bdef38cd52" Nov 24 07:13:52 crc kubenswrapper[4809]: E1124 07:13:52.940054 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"110725378d5569ef98bf845f63c45738e84831d095aa4ac8049621bdef38cd52\": container with ID starting with 110725378d5569ef98bf845f63c45738e84831d095aa4ac8049621bdef38cd52 not found: ID does not exist" containerID="110725378d5569ef98bf845f63c45738e84831d095aa4ac8049621bdef38cd52" Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.940094 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"110725378d5569ef98bf845f63c45738e84831d095aa4ac8049621bdef38cd52"} err="failed to get container status \"110725378d5569ef98bf845f63c45738e84831d095aa4ac8049621bdef38cd52\": rpc error: code = NotFound desc = could not find container \"110725378d5569ef98bf845f63c45738e84831d095aa4ac8049621bdef38cd52\": container with ID starting with 110725378d5569ef98bf845f63c45738e84831d095aa4ac8049621bdef38cd52 not found: ID does not exist" Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.966909 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-dns-swift-storage-0\") pod \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.966973 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-dns-svc\") pod \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.967014 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-config\") pod \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.967035 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-ovsdbserver-sb\") pod \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.967063 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-ovsdbserver-nb\") pod \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.967181 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtbh2\" (UniqueName: \"kubernetes.io/projected/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-kube-api-access-wtbh2\") pod \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\" (UID: \"ec7ac87d-2d70-4384-aceb-6f2eca6bf94b\") " Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.974921 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-kube-api-access-wtbh2" (OuterVolumeSpecName: "kube-api-access-wtbh2") pod "ec7ac87d-2d70-4384-aceb-6f2eca6bf94b" (UID: "ec7ac87d-2d70-4384-aceb-6f2eca6bf94b"). InnerVolumeSpecName "kube-api-access-wtbh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:13:52 crc kubenswrapper[4809]: I1124 07:13:52.987496 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.042650 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ec7ac87d-2d70-4384-aceb-6f2eca6bf94b" (UID: "ec7ac87d-2d70-4384-aceb-6f2eca6bf94b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.044987 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ec7ac87d-2d70-4384-aceb-6f2eca6bf94b" (UID: "ec7ac87d-2d70-4384-aceb-6f2eca6bf94b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.070004 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56d958c1-a749-4f2f-b9a7-d5d526da77c7-config-data\") pod \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\" (UID: \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\") " Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.070610 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56d958c1-a749-4f2f-b9a7-d5d526da77c7-logs\") pod \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\" (UID: \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\") " Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.070702 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkz2z\" (UniqueName: \"kubernetes.io/projected/56d958c1-a749-4f2f-b9a7-d5d526da77c7-kube-api-access-zkz2z\") pod \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\" (UID: \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\") " Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.070830 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/56d958c1-a749-4f2f-b9a7-d5d526da77c7-nova-metadata-tls-certs\") pod \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\" (UID: \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\") " Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.070918 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56d958c1-a749-4f2f-b9a7-d5d526da77c7-combined-ca-bundle\") pod \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\" (UID: \"56d958c1-a749-4f2f-b9a7-d5d526da77c7\") " Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.071422 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtbh2\" (UniqueName: \"kubernetes.io/projected/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-kube-api-access-wtbh2\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.071494 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.071550 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.074450 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ec7ac87d-2d70-4384-aceb-6f2eca6bf94b" (UID: "ec7ac87d-2d70-4384-aceb-6f2eca6bf94b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.075951 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56d958c1-a749-4f2f-b9a7-d5d526da77c7-kube-api-access-zkz2z" (OuterVolumeSpecName: "kube-api-access-zkz2z") pod "56d958c1-a749-4f2f-b9a7-d5d526da77c7" (UID: "56d958c1-a749-4f2f-b9a7-d5d526da77c7"). InnerVolumeSpecName "kube-api-access-zkz2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.076179 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56d958c1-a749-4f2f-b9a7-d5d526da77c7-logs" (OuterVolumeSpecName: "logs") pod "56d958c1-a749-4f2f-b9a7-d5d526da77c7" (UID: "56d958c1-a749-4f2f-b9a7-d5d526da77c7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.076379 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.081478 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-config" (OuterVolumeSpecName: "config") pod "ec7ac87d-2d70-4384-aceb-6f2eca6bf94b" (UID: "ec7ac87d-2d70-4384-aceb-6f2eca6bf94b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.088190 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ec7ac87d-2d70-4384-aceb-6f2eca6bf94b" (UID: "ec7ac87d-2d70-4384-aceb-6f2eca6bf94b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.113784 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56d958c1-a749-4f2f-b9a7-d5d526da77c7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56d958c1-a749-4f2f-b9a7-d5d526da77c7" (UID: "56d958c1-a749-4f2f-b9a7-d5d526da77c7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.132141 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56d958c1-a749-4f2f-b9a7-d5d526da77c7-config-data" (OuterVolumeSpecName: "config-data") pod "56d958c1-a749-4f2f-b9a7-d5d526da77c7" (UID: "56d958c1-a749-4f2f-b9a7-d5d526da77c7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.141688 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56d958c1-a749-4f2f-b9a7-d5d526da77c7-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "56d958c1-a749-4f2f-b9a7-d5d526da77c7" (UID: "56d958c1-a749-4f2f-b9a7-d5d526da77c7"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.173358 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-combined-ca-bundle\") pod \"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c\" (UID: \"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c\") " Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.173620 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-logs\") pod \"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c\" (UID: \"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c\") " Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.173725 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bxqb\" (UniqueName: \"kubernetes.io/projected/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-kube-api-access-9bxqb\") pod \"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c\" (UID: \"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c\") " Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.173904 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-config-data\") pod \"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c\" (UID: \"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c\") " Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.174002 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-logs" (OuterVolumeSpecName: "logs") pod "b71bf545-2e9d-4dd4-bd8a-db1eb36a535c" (UID: "b71bf545-2e9d-4dd4-bd8a-db1eb36a535c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.174402 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56d958c1-a749-4f2f-b9a7-d5d526da77c7-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.174467 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.174590 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.174646 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56d958c1-a749-4f2f-b9a7-d5d526da77c7-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.174704 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.174758 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.174812 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkz2z\" (UniqueName: \"kubernetes.io/projected/56d958c1-a749-4f2f-b9a7-d5d526da77c7-kube-api-access-zkz2z\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.175234 4809 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/56d958c1-a749-4f2f-b9a7-d5d526da77c7-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.175302 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56d958c1-a749-4f2f-b9a7-d5d526da77c7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.177861 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-kube-api-access-9bxqb" (OuterVolumeSpecName: "kube-api-access-9bxqb") pod "b71bf545-2e9d-4dd4-bd8a-db1eb36a535c" (UID: "b71bf545-2e9d-4dd4-bd8a-db1eb36a535c"). InnerVolumeSpecName "kube-api-access-9bxqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.213574 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-config-data" (OuterVolumeSpecName: "config-data") pod "b71bf545-2e9d-4dd4-bd8a-db1eb36a535c" (UID: "b71bf545-2e9d-4dd4-bd8a-db1eb36a535c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.216351 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b71bf545-2e9d-4dd4-bd8a-db1eb36a535c" (UID: "b71bf545-2e9d-4dd4-bd8a-db1eb36a535c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.276865 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bxqb\" (UniqueName: \"kubernetes.io/projected/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-kube-api-access-9bxqb\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.276893 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.276902 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.285201 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9fqbw" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.295599 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-br547"] Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.305009 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-br547"] Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.378196 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45003542-95f6-46b1-8ff9-06dfc99c1d93-config-data\") pod \"45003542-95f6-46b1-8ff9-06dfc99c1d93\" (UID: \"45003542-95f6-46b1-8ff9-06dfc99c1d93\") " Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.378287 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45003542-95f6-46b1-8ff9-06dfc99c1d93-scripts\") pod \"45003542-95f6-46b1-8ff9-06dfc99c1d93\" (UID: \"45003542-95f6-46b1-8ff9-06dfc99c1d93\") " Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.378330 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45003542-95f6-46b1-8ff9-06dfc99c1d93-combined-ca-bundle\") pod \"45003542-95f6-46b1-8ff9-06dfc99c1d93\" (UID: \"45003542-95f6-46b1-8ff9-06dfc99c1d93\") " Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.378415 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsxwr\" (UniqueName: \"kubernetes.io/projected/45003542-95f6-46b1-8ff9-06dfc99c1d93-kube-api-access-hsxwr\") pod \"45003542-95f6-46b1-8ff9-06dfc99c1d93\" (UID: \"45003542-95f6-46b1-8ff9-06dfc99c1d93\") " Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.382025 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45003542-95f6-46b1-8ff9-06dfc99c1d93-kube-api-access-hsxwr" (OuterVolumeSpecName: "kube-api-access-hsxwr") pod "45003542-95f6-46b1-8ff9-06dfc99c1d93" (UID: "45003542-95f6-46b1-8ff9-06dfc99c1d93"). InnerVolumeSpecName "kube-api-access-hsxwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.383047 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45003542-95f6-46b1-8ff9-06dfc99c1d93-scripts" (OuterVolumeSpecName: "scripts") pod "45003542-95f6-46b1-8ff9-06dfc99c1d93" (UID: "45003542-95f6-46b1-8ff9-06dfc99c1d93"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.406495 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45003542-95f6-46b1-8ff9-06dfc99c1d93-config-data" (OuterVolumeSpecName: "config-data") pod "45003542-95f6-46b1-8ff9-06dfc99c1d93" (UID: "45003542-95f6-46b1-8ff9-06dfc99c1d93"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.422239 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45003542-95f6-46b1-8ff9-06dfc99c1d93-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "45003542-95f6-46b1-8ff9-06dfc99c1d93" (UID: "45003542-95f6-46b1-8ff9-06dfc99c1d93"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.480689 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45003542-95f6-46b1-8ff9-06dfc99c1d93-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.480740 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45003542-95f6-46b1-8ff9-06dfc99c1d93-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.480761 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45003542-95f6-46b1-8ff9-06dfc99c1d93-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.480782 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsxwr\" (UniqueName: \"kubernetes.io/projected/45003542-95f6-46b1-8ff9-06dfc99c1d93-kube-api-access-hsxwr\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.894038 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b71bf545-2e9d-4dd4-bd8a-db1eb36a535c","Type":"ContainerDied","Data":"bf7389cb17017f91f0e763fb12ce005195c09ccd9aba4abcfadae346ab3a4a04"} Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.894103 4809 scope.go:117] "RemoveContainer" containerID="af6567b44f00bee9c3dd572b507523c134d390af573d5d4989dd6ce9c78b9065" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.894228 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.910205 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9fqbw" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.910950 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9fqbw" event={"ID":"45003542-95f6-46b1-8ff9-06dfc99c1d93","Type":"ContainerDied","Data":"34c81e6a807fa92cda14c6ec661a3150cdfc4d96fbcd41246a4d82955a961820"} Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.911007 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34c81e6a807fa92cda14c6ec661a3150cdfc4d96fbcd41246a4d82955a961820" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.917597 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="fb60935a-17c8-445c-b4ab-99786f623acb" containerName="nova-scheduler-scheduler" containerID="cri-o://5dc457b232c378d3ca32c708e2c0773ddd5918ba4bdd32c6f2ada08f52dc3b84" gracePeriod=30 Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.917860 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.920069 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"56d958c1-a749-4f2f-b9a7-d5d526da77c7","Type":"ContainerDied","Data":"9bcd4beda3f2dfd786a4d5ef4e1aa85e3b26da7fd8508e7b0ef15f63db470dc1"} Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.977018 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 07:13:53 crc kubenswrapper[4809]: E1124 07:13:53.977511 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec7ac87d-2d70-4384-aceb-6f2eca6bf94b" containerName="init" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.977530 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec7ac87d-2d70-4384-aceb-6f2eca6bf94b" containerName="init" Nov 24 07:13:53 crc kubenswrapper[4809]: E1124 07:13:53.977538 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56d958c1-a749-4f2f-b9a7-d5d526da77c7" containerName="nova-metadata-metadata" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.977546 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="56d958c1-a749-4f2f-b9a7-d5d526da77c7" containerName="nova-metadata-metadata" Nov 24 07:13:53 crc kubenswrapper[4809]: E1124 07:13:53.977569 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec7ac87d-2d70-4384-aceb-6f2eca6bf94b" containerName="dnsmasq-dns" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.977577 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec7ac87d-2d70-4384-aceb-6f2eca6bf94b" containerName="dnsmasq-dns" Nov 24 07:13:53 crc kubenswrapper[4809]: E1124 07:13:53.977589 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b71bf545-2e9d-4dd4-bd8a-db1eb36a535c" containerName="nova-api-api" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.977597 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b71bf545-2e9d-4dd4-bd8a-db1eb36a535c" containerName="nova-api-api" Nov 24 07:13:53 crc kubenswrapper[4809]: E1124 07:13:53.977620 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b71bf545-2e9d-4dd4-bd8a-db1eb36a535c" containerName="nova-api-log" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.977628 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b71bf545-2e9d-4dd4-bd8a-db1eb36a535c" containerName="nova-api-log" Nov 24 07:13:53 crc kubenswrapper[4809]: E1124 07:13:53.977646 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45003542-95f6-46b1-8ff9-06dfc99c1d93" containerName="nova-cell1-conductor-db-sync" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.977654 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="45003542-95f6-46b1-8ff9-06dfc99c1d93" containerName="nova-cell1-conductor-db-sync" Nov 24 07:13:53 crc kubenswrapper[4809]: E1124 07:13:53.977677 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56d958c1-a749-4f2f-b9a7-d5d526da77c7" containerName="nova-metadata-log" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.977686 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="56d958c1-a749-4f2f-b9a7-d5d526da77c7" containerName="nova-metadata-log" Nov 24 07:13:53 crc kubenswrapper[4809]: E1124 07:13:53.977707 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="825d5761-5b87-4e12-8e7c-392e252331e0" containerName="nova-manage" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.977714 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="825d5761-5b87-4e12-8e7c-392e252331e0" containerName="nova-manage" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.977926 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="56d958c1-a749-4f2f-b9a7-d5d526da77c7" containerName="nova-metadata-log" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.977947 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b71bf545-2e9d-4dd4-bd8a-db1eb36a535c" containerName="nova-api-log" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.980101 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec7ac87d-2d70-4384-aceb-6f2eca6bf94b" containerName="dnsmasq-dns" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.980137 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="45003542-95f6-46b1-8ff9-06dfc99c1d93" containerName="nova-cell1-conductor-db-sync" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.980154 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="825d5761-5b87-4e12-8e7c-392e252331e0" containerName="nova-manage" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.980166 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b71bf545-2e9d-4dd4-bd8a-db1eb36a535c" containerName="nova-api-api" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.980189 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="56d958c1-a749-4f2f-b9a7-d5d526da77c7" containerName="nova-metadata-metadata" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.980459 4809 scope.go:117] "RemoveContainer" containerID="09111d790ecf2eed86fbfcc0c474c956ffdc7d6a7682d190d63d8ba5cc12e8d6" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.981037 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.985826 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 07:13:53 crc kubenswrapper[4809]: I1124 07:13:53.998078 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.014624 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.037230 4809 scope.go:117] "RemoveContainer" containerID="4a90f038dc639efc80174ba98a119140ecc2f2acc75f680554a55f967ac9ac23" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.048631 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.069137 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.075267 4809 scope.go:117] "RemoveContainer" containerID="5e4fa1f28f7adf3189d1f81512b329afc3e9694a9210ce4734632959889f2840" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.081088 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.093675 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.097197 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.098842 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.100127 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g96rm\" (UniqueName: \"kubernetes.io/projected/f4318b40-eb3b-4de1-8c8f-c9e4bf5996f8-kube-api-access-g96rm\") pod \"nova-cell1-conductor-0\" (UID: \"f4318b40-eb3b-4de1-8c8f-c9e4bf5996f8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.100200 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4318b40-eb3b-4de1-8c8f-c9e4bf5996f8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f4318b40-eb3b-4de1-8c8f-c9e4bf5996f8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.100346 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4318b40-eb3b-4de1-8c8f-c9e4bf5996f8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f4318b40-eb3b-4de1-8c8f-c9e4bf5996f8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.101256 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.118065 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.119645 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.121779 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.122297 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.129689 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.202839 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c0203aa-9e75-4bf4-8660-68f27b1b794b-logs\") pod \"nova-api-0\" (UID: \"2c0203aa-9e75-4bf4-8660-68f27b1b794b\") " pod="openstack/nova-api-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.202919 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4318b40-eb3b-4de1-8c8f-c9e4bf5996f8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f4318b40-eb3b-4de1-8c8f-c9e4bf5996f8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.202947 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c0203aa-9e75-4bf4-8660-68f27b1b794b-config-data\") pod \"nova-api-0\" (UID: \"2c0203aa-9e75-4bf4-8660-68f27b1b794b\") " pod="openstack/nova-api-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.203020 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4318b40-eb3b-4de1-8c8f-c9e4bf5996f8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f4318b40-eb3b-4de1-8c8f-c9e4bf5996f8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.203048 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvgkr\" (UniqueName: \"kubernetes.io/projected/2c0203aa-9e75-4bf4-8660-68f27b1b794b-kube-api-access-zvgkr\") pod \"nova-api-0\" (UID: \"2c0203aa-9e75-4bf4-8660-68f27b1b794b\") " pod="openstack/nova-api-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.203093 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c0203aa-9e75-4bf4-8660-68f27b1b794b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2c0203aa-9e75-4bf4-8660-68f27b1b794b\") " pod="openstack/nova-api-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.203194 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g96rm\" (UniqueName: \"kubernetes.io/projected/f4318b40-eb3b-4de1-8c8f-c9e4bf5996f8-kube-api-access-g96rm\") pod \"nova-cell1-conductor-0\" (UID: \"f4318b40-eb3b-4de1-8c8f-c9e4bf5996f8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.207345 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4318b40-eb3b-4de1-8c8f-c9e4bf5996f8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f4318b40-eb3b-4de1-8c8f-c9e4bf5996f8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.208604 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4318b40-eb3b-4de1-8c8f-c9e4bf5996f8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f4318b40-eb3b-4de1-8c8f-c9e4bf5996f8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.219807 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g96rm\" (UniqueName: \"kubernetes.io/projected/f4318b40-eb3b-4de1-8c8f-c9e4bf5996f8-kube-api-access-g96rm\") pod \"nova-cell1-conductor-0\" (UID: \"f4318b40-eb3b-4de1-8c8f-c9e4bf5996f8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.305065 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1286d658-6187-4509-9e24-1e0d6fbcab1e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1286d658-6187-4509-9e24-1e0d6fbcab1e\") " pod="openstack/nova-metadata-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.305116 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1286d658-6187-4509-9e24-1e0d6fbcab1e-config-data\") pod \"nova-metadata-0\" (UID: \"1286d658-6187-4509-9e24-1e0d6fbcab1e\") " pod="openstack/nova-metadata-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.305162 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1286d658-6187-4509-9e24-1e0d6fbcab1e-logs\") pod \"nova-metadata-0\" (UID: \"1286d658-6187-4509-9e24-1e0d6fbcab1e\") " pod="openstack/nova-metadata-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.305191 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c0203aa-9e75-4bf4-8660-68f27b1b794b-logs\") pod \"nova-api-0\" (UID: \"2c0203aa-9e75-4bf4-8660-68f27b1b794b\") " pod="openstack/nova-api-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.305208 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztbps\" (UniqueName: \"kubernetes.io/projected/1286d658-6187-4509-9e24-1e0d6fbcab1e-kube-api-access-ztbps\") pod \"nova-metadata-0\" (UID: \"1286d658-6187-4509-9e24-1e0d6fbcab1e\") " pod="openstack/nova-metadata-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.305243 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c0203aa-9e75-4bf4-8660-68f27b1b794b-config-data\") pod \"nova-api-0\" (UID: \"2c0203aa-9e75-4bf4-8660-68f27b1b794b\") " pod="openstack/nova-api-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.305266 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1286d658-6187-4509-9e24-1e0d6fbcab1e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1286d658-6187-4509-9e24-1e0d6fbcab1e\") " pod="openstack/nova-metadata-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.305410 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvgkr\" (UniqueName: \"kubernetes.io/projected/2c0203aa-9e75-4bf4-8660-68f27b1b794b-kube-api-access-zvgkr\") pod \"nova-api-0\" (UID: \"2c0203aa-9e75-4bf4-8660-68f27b1b794b\") " pod="openstack/nova-api-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.305482 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c0203aa-9e75-4bf4-8660-68f27b1b794b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2c0203aa-9e75-4bf4-8660-68f27b1b794b\") " pod="openstack/nova-api-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.306191 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c0203aa-9e75-4bf4-8660-68f27b1b794b-logs\") pod \"nova-api-0\" (UID: \"2c0203aa-9e75-4bf4-8660-68f27b1b794b\") " pod="openstack/nova-api-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.309500 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c0203aa-9e75-4bf4-8660-68f27b1b794b-config-data\") pod \"nova-api-0\" (UID: \"2c0203aa-9e75-4bf4-8660-68f27b1b794b\") " pod="openstack/nova-api-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.321911 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.322164 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c0203aa-9e75-4bf4-8660-68f27b1b794b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2c0203aa-9e75-4bf4-8660-68f27b1b794b\") " pod="openstack/nova-api-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.327251 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvgkr\" (UniqueName: \"kubernetes.io/projected/2c0203aa-9e75-4bf4-8660-68f27b1b794b-kube-api-access-zvgkr\") pod \"nova-api-0\" (UID: \"2c0203aa-9e75-4bf4-8660-68f27b1b794b\") " pod="openstack/nova-api-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.407646 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1286d658-6187-4509-9e24-1e0d6fbcab1e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1286d658-6187-4509-9e24-1e0d6fbcab1e\") " pod="openstack/nova-metadata-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.407797 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1286d658-6187-4509-9e24-1e0d6fbcab1e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1286d658-6187-4509-9e24-1e0d6fbcab1e\") " pod="openstack/nova-metadata-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.407826 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1286d658-6187-4509-9e24-1e0d6fbcab1e-config-data\") pod \"nova-metadata-0\" (UID: \"1286d658-6187-4509-9e24-1e0d6fbcab1e\") " pod="openstack/nova-metadata-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.407882 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1286d658-6187-4509-9e24-1e0d6fbcab1e-logs\") pod \"nova-metadata-0\" (UID: \"1286d658-6187-4509-9e24-1e0d6fbcab1e\") " pod="openstack/nova-metadata-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.407917 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztbps\" (UniqueName: \"kubernetes.io/projected/1286d658-6187-4509-9e24-1e0d6fbcab1e-kube-api-access-ztbps\") pod \"nova-metadata-0\" (UID: \"1286d658-6187-4509-9e24-1e0d6fbcab1e\") " pod="openstack/nova-metadata-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.412729 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1286d658-6187-4509-9e24-1e0d6fbcab1e-logs\") pod \"nova-metadata-0\" (UID: \"1286d658-6187-4509-9e24-1e0d6fbcab1e\") " pod="openstack/nova-metadata-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.414276 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1286d658-6187-4509-9e24-1e0d6fbcab1e-config-data\") pod \"nova-metadata-0\" (UID: \"1286d658-6187-4509-9e24-1e0d6fbcab1e\") " pod="openstack/nova-metadata-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.415214 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1286d658-6187-4509-9e24-1e0d6fbcab1e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1286d658-6187-4509-9e24-1e0d6fbcab1e\") " pod="openstack/nova-metadata-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.417621 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.421292 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1286d658-6187-4509-9e24-1e0d6fbcab1e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1286d658-6187-4509-9e24-1e0d6fbcab1e\") " pod="openstack/nova-metadata-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.428627 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztbps\" (UniqueName: \"kubernetes.io/projected/1286d658-6187-4509-9e24-1e0d6fbcab1e-kube-api-access-ztbps\") pod \"nova-metadata-0\" (UID: \"1286d658-6187-4509-9e24-1e0d6fbcab1e\") " pod="openstack/nova-metadata-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.440540 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.766633 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 07:13:54 crc kubenswrapper[4809]: W1124 07:13:54.778164 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4318b40_eb3b_4de1_8c8f_c9e4bf5996f8.slice/crio-0e85ba4a8801b7171c8e9ff9ace3dc9c8fd6bde8650f3ab4d052d1cb2db94cdf WatchSource:0}: Error finding container 0e85ba4a8801b7171c8e9ff9ace3dc9c8fd6bde8650f3ab4d052d1cb2db94cdf: Status 404 returned error can't find the container with id 0e85ba4a8801b7171c8e9ff9ace3dc9c8fd6bde8650f3ab4d052d1cb2db94cdf Nov 24 07:13:54 crc kubenswrapper[4809]: W1124 07:13:54.788587 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1286d658_6187_4509_9e24_1e0d6fbcab1e.slice/crio-9e45992ea7bcf9d34a0895b715515999f4e06ab76942080ec1f9657fb87776c4 WatchSource:0}: Error finding container 9e45992ea7bcf9d34a0895b715515999f4e06ab76942080ec1f9657fb87776c4: Status 404 returned error can't find the container with id 9e45992ea7bcf9d34a0895b715515999f4e06ab76942080ec1f9657fb87776c4 Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.788979 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.909942 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56d958c1-a749-4f2f-b9a7-d5d526da77c7" path="/var/lib/kubelet/pods/56d958c1-a749-4f2f-b9a7-d5d526da77c7/volumes" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.911390 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b71bf545-2e9d-4dd4-bd8a-db1eb36a535c" path="/var/lib/kubelet/pods/b71bf545-2e9d-4dd4-bd8a-db1eb36a535c/volumes" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.912180 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec7ac87d-2d70-4384-aceb-6f2eca6bf94b" path="/var/lib/kubelet/pods/ec7ac87d-2d70-4384-aceb-6f2eca6bf94b/volumes" Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.915020 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:13:54 crc kubenswrapper[4809]: W1124 07:13:54.925941 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c0203aa_9e75_4bf4_8660_68f27b1b794b.slice/crio-684bc7f2cbd1c89af75e2704aa5f285e26d138c82a16653bf1a4bf90b9cdf5bc WatchSource:0}: Error finding container 684bc7f2cbd1c89af75e2704aa5f285e26d138c82a16653bf1a4bf90b9cdf5bc: Status 404 returned error can't find the container with id 684bc7f2cbd1c89af75e2704aa5f285e26d138c82a16653bf1a4bf90b9cdf5bc Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.929855 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"f4318b40-eb3b-4de1-8c8f-c9e4bf5996f8","Type":"ContainerStarted","Data":"0e85ba4a8801b7171c8e9ff9ace3dc9c8fd6bde8650f3ab4d052d1cb2db94cdf"} Nov 24 07:13:54 crc kubenswrapper[4809]: I1124 07:13:54.932682 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1286d658-6187-4509-9e24-1e0d6fbcab1e","Type":"ContainerStarted","Data":"9e45992ea7bcf9d34a0895b715515999f4e06ab76942080ec1f9657fb87776c4"} Nov 24 07:13:55 crc kubenswrapper[4809]: I1124 07:13:55.947258 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2c0203aa-9e75-4bf4-8660-68f27b1b794b","Type":"ContainerStarted","Data":"a8c9a69b328f2e57ec6455cd0c4ee70ec9c896be31e4103c6cac1edb7d7381ca"} Nov 24 07:13:55 crc kubenswrapper[4809]: I1124 07:13:55.947318 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2c0203aa-9e75-4bf4-8660-68f27b1b794b","Type":"ContainerStarted","Data":"19bf512267a342fe3d48842756d71bdb73d4717364dfec75d2e3e6d89ef9578f"} Nov 24 07:13:55 crc kubenswrapper[4809]: I1124 07:13:55.947339 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2c0203aa-9e75-4bf4-8660-68f27b1b794b","Type":"ContainerStarted","Data":"684bc7f2cbd1c89af75e2704aa5f285e26d138c82a16653bf1a4bf90b9cdf5bc"} Nov 24 07:13:55 crc kubenswrapper[4809]: I1124 07:13:55.949629 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1286d658-6187-4509-9e24-1e0d6fbcab1e","Type":"ContainerStarted","Data":"8252b5e689178dcecb351ad3aeeb9896ac6dbaa4ca0ff97c73985ad51a08b07b"} Nov 24 07:13:55 crc kubenswrapper[4809]: I1124 07:13:55.949686 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1286d658-6187-4509-9e24-1e0d6fbcab1e","Type":"ContainerStarted","Data":"0a3e1f620cd8d80a95f532ec6d9f8cb9f3a36e6385c2d51b423c6d336e475487"} Nov 24 07:13:55 crc kubenswrapper[4809]: I1124 07:13:55.951528 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"f4318b40-eb3b-4de1-8c8f-c9e4bf5996f8","Type":"ContainerStarted","Data":"95966e1a59cc916d0558f116d757b0f05b5bc9997bf4f1404b6bf2865346e45f"} Nov 24 07:13:55 crc kubenswrapper[4809]: I1124 07:13:55.951753 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 24 07:13:55 crc kubenswrapper[4809]: I1124 07:13:55.970989 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.970951745 podStartE2EDuration="2.970951745s" podCreationTimestamp="2025-11-24 07:13:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:13:55.96399947 +0000 UTC m=+1135.864591185" watchObservedRunningTime="2025-11-24 07:13:55.970951745 +0000 UTC m=+1135.871543460" Nov 24 07:13:55 crc kubenswrapper[4809]: I1124 07:13:55.994829 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.994806577 podStartE2EDuration="2.994806577s" podCreationTimestamp="2025-11-24 07:13:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:13:55.984881374 +0000 UTC m=+1135.885473089" watchObservedRunningTime="2025-11-24 07:13:55.994806577 +0000 UTC m=+1135.895398292" Nov 24 07:13:56 crc kubenswrapper[4809]: I1124 07:13:56.017025 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.017000275 podStartE2EDuration="3.017000275s" podCreationTimestamp="2025-11-24 07:13:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:13:56.012928608 +0000 UTC m=+1135.913520373" watchObservedRunningTime="2025-11-24 07:13:56.017000275 +0000 UTC m=+1135.917591990" Nov 24 07:13:56 crc kubenswrapper[4809]: I1124 07:13:56.966044 4809 generic.go:334] "Generic (PLEG): container finished" podID="fb60935a-17c8-445c-b4ab-99786f623acb" containerID="5dc457b232c378d3ca32c708e2c0773ddd5918ba4bdd32c6f2ada08f52dc3b84" exitCode=0 Nov 24 07:13:56 crc kubenswrapper[4809]: I1124 07:13:56.966168 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fb60935a-17c8-445c-b4ab-99786f623acb","Type":"ContainerDied","Data":"5dc457b232c378d3ca32c708e2c0773ddd5918ba4bdd32c6f2ada08f52dc3b84"} Nov 24 07:13:56 crc kubenswrapper[4809]: E1124 07:13:56.967633 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5dc457b232c378d3ca32c708e2c0773ddd5918ba4bdd32c6f2ada08f52dc3b84 is running failed: container process not found" containerID="5dc457b232c378d3ca32c708e2c0773ddd5918ba4bdd32c6f2ada08f52dc3b84" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 07:13:56 crc kubenswrapper[4809]: E1124 07:13:56.968210 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5dc457b232c378d3ca32c708e2c0773ddd5918ba4bdd32c6f2ada08f52dc3b84 is running failed: container process not found" containerID="5dc457b232c378d3ca32c708e2c0773ddd5918ba4bdd32c6f2ada08f52dc3b84" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 07:13:56 crc kubenswrapper[4809]: E1124 07:13:56.968753 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5dc457b232c378d3ca32c708e2c0773ddd5918ba4bdd32c6f2ada08f52dc3b84 is running failed: container process not found" containerID="5dc457b232c378d3ca32c708e2c0773ddd5918ba4bdd32c6f2ada08f52dc3b84" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 07:13:56 crc kubenswrapper[4809]: E1124 07:13:56.968833 4809 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5dc457b232c378d3ca32c708e2c0773ddd5918ba4bdd32c6f2ada08f52dc3b84 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="fb60935a-17c8-445c-b4ab-99786f623acb" containerName="nova-scheduler-scheduler" Nov 24 07:13:57 crc kubenswrapper[4809]: I1124 07:13:57.303013 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 07:13:57 crc kubenswrapper[4809]: I1124 07:13:57.370797 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb60935a-17c8-445c-b4ab-99786f623acb-config-data\") pod \"fb60935a-17c8-445c-b4ab-99786f623acb\" (UID: \"fb60935a-17c8-445c-b4ab-99786f623acb\") " Nov 24 07:13:57 crc kubenswrapper[4809]: I1124 07:13:57.371031 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-777kw\" (UniqueName: \"kubernetes.io/projected/fb60935a-17c8-445c-b4ab-99786f623acb-kube-api-access-777kw\") pod \"fb60935a-17c8-445c-b4ab-99786f623acb\" (UID: \"fb60935a-17c8-445c-b4ab-99786f623acb\") " Nov 24 07:13:57 crc kubenswrapper[4809]: I1124 07:13:57.371148 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb60935a-17c8-445c-b4ab-99786f623acb-combined-ca-bundle\") pod \"fb60935a-17c8-445c-b4ab-99786f623acb\" (UID: \"fb60935a-17c8-445c-b4ab-99786f623acb\") " Nov 24 07:13:57 crc kubenswrapper[4809]: I1124 07:13:57.378343 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb60935a-17c8-445c-b4ab-99786f623acb-kube-api-access-777kw" (OuterVolumeSpecName: "kube-api-access-777kw") pod "fb60935a-17c8-445c-b4ab-99786f623acb" (UID: "fb60935a-17c8-445c-b4ab-99786f623acb"). InnerVolumeSpecName "kube-api-access-777kw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:13:57 crc kubenswrapper[4809]: I1124 07:13:57.400734 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb60935a-17c8-445c-b4ab-99786f623acb-config-data" (OuterVolumeSpecName: "config-data") pod "fb60935a-17c8-445c-b4ab-99786f623acb" (UID: "fb60935a-17c8-445c-b4ab-99786f623acb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:57 crc kubenswrapper[4809]: I1124 07:13:57.407418 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb60935a-17c8-445c-b4ab-99786f623acb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb60935a-17c8-445c-b4ab-99786f623acb" (UID: "fb60935a-17c8-445c-b4ab-99786f623acb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:13:57 crc kubenswrapper[4809]: I1124 07:13:57.472902 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-777kw\" (UniqueName: \"kubernetes.io/projected/fb60935a-17c8-445c-b4ab-99786f623acb-kube-api-access-777kw\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:57 crc kubenswrapper[4809]: I1124 07:13:57.472939 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb60935a-17c8-445c-b4ab-99786f623acb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:57 crc kubenswrapper[4809]: I1124 07:13:57.472951 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb60935a-17c8-445c-b4ab-99786f623acb-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:13:57 crc kubenswrapper[4809]: I1124 07:13:57.900620 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 07:13:58 crc kubenswrapper[4809]: I1124 07:13:58.000184 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fb60935a-17c8-445c-b4ab-99786f623acb","Type":"ContainerDied","Data":"7f04a43a86e56e5ae3244d709a800ca90776dda5a055ef4469923707ab5cf68f"} Nov 24 07:13:58 crc kubenswrapper[4809]: I1124 07:13:58.000237 4809 scope.go:117] "RemoveContainer" containerID="5dc457b232c378d3ca32c708e2c0773ddd5918ba4bdd32c6f2ada08f52dc3b84" Nov 24 07:13:58 crc kubenswrapper[4809]: I1124 07:13:58.000251 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 07:13:58 crc kubenswrapper[4809]: I1124 07:13:58.048600 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:13:58 crc kubenswrapper[4809]: I1124 07:13:58.074021 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:13:58 crc kubenswrapper[4809]: I1124 07:13:58.089463 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:13:58 crc kubenswrapper[4809]: E1124 07:13:58.089953 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb60935a-17c8-445c-b4ab-99786f623acb" containerName="nova-scheduler-scheduler" Nov 24 07:13:58 crc kubenswrapper[4809]: I1124 07:13:58.089997 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb60935a-17c8-445c-b4ab-99786f623acb" containerName="nova-scheduler-scheduler" Nov 24 07:13:58 crc kubenswrapper[4809]: I1124 07:13:58.090255 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb60935a-17c8-445c-b4ab-99786f623acb" containerName="nova-scheduler-scheduler" Nov 24 07:13:58 crc kubenswrapper[4809]: I1124 07:13:58.091311 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 07:13:58 crc kubenswrapper[4809]: I1124 07:13:58.093360 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 07:13:58 crc kubenswrapper[4809]: I1124 07:13:58.101758 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:13:58 crc kubenswrapper[4809]: I1124 07:13:58.191474 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw4gs\" (UniqueName: \"kubernetes.io/projected/111098ed-9d16-45e0-bb80-35f9383e327e-kube-api-access-cw4gs\") pod \"nova-scheduler-0\" (UID: \"111098ed-9d16-45e0-bb80-35f9383e327e\") " pod="openstack/nova-scheduler-0" Nov 24 07:13:58 crc kubenswrapper[4809]: I1124 07:13:58.191790 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/111098ed-9d16-45e0-bb80-35f9383e327e-config-data\") pod \"nova-scheduler-0\" (UID: \"111098ed-9d16-45e0-bb80-35f9383e327e\") " pod="openstack/nova-scheduler-0" Nov 24 07:13:58 crc kubenswrapper[4809]: I1124 07:13:58.192335 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/111098ed-9d16-45e0-bb80-35f9383e327e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"111098ed-9d16-45e0-bb80-35f9383e327e\") " pod="openstack/nova-scheduler-0" Nov 24 07:13:58 crc kubenswrapper[4809]: I1124 07:13:58.294219 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/111098ed-9d16-45e0-bb80-35f9383e327e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"111098ed-9d16-45e0-bb80-35f9383e327e\") " pod="openstack/nova-scheduler-0" Nov 24 07:13:58 crc kubenswrapper[4809]: I1124 07:13:58.294302 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw4gs\" (UniqueName: \"kubernetes.io/projected/111098ed-9d16-45e0-bb80-35f9383e327e-kube-api-access-cw4gs\") pod \"nova-scheduler-0\" (UID: \"111098ed-9d16-45e0-bb80-35f9383e327e\") " pod="openstack/nova-scheduler-0" Nov 24 07:13:58 crc kubenswrapper[4809]: I1124 07:13:58.294325 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/111098ed-9d16-45e0-bb80-35f9383e327e-config-data\") pod \"nova-scheduler-0\" (UID: \"111098ed-9d16-45e0-bb80-35f9383e327e\") " pod="openstack/nova-scheduler-0" Nov 24 07:13:58 crc kubenswrapper[4809]: I1124 07:13:58.299043 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/111098ed-9d16-45e0-bb80-35f9383e327e-config-data\") pod \"nova-scheduler-0\" (UID: \"111098ed-9d16-45e0-bb80-35f9383e327e\") " pod="openstack/nova-scheduler-0" Nov 24 07:13:58 crc kubenswrapper[4809]: I1124 07:13:58.299804 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/111098ed-9d16-45e0-bb80-35f9383e327e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"111098ed-9d16-45e0-bb80-35f9383e327e\") " pod="openstack/nova-scheduler-0" Nov 24 07:13:58 crc kubenswrapper[4809]: I1124 07:13:58.312555 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw4gs\" (UniqueName: \"kubernetes.io/projected/111098ed-9d16-45e0-bb80-35f9383e327e-kube-api-access-cw4gs\") pod \"nova-scheduler-0\" (UID: \"111098ed-9d16-45e0-bb80-35f9383e327e\") " pod="openstack/nova-scheduler-0" Nov 24 07:13:58 crc kubenswrapper[4809]: I1124 07:13:58.417852 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 07:13:58 crc kubenswrapper[4809]: W1124 07:13:58.855716 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod111098ed_9d16_45e0_bb80_35f9383e327e.slice/crio-276bdc3c5eb76ef9fadb7fc2fe909c352901ea24ae7bade45885aaa3a2991ef6 WatchSource:0}: Error finding container 276bdc3c5eb76ef9fadb7fc2fe909c352901ea24ae7bade45885aaa3a2991ef6: Status 404 returned error can't find the container with id 276bdc3c5eb76ef9fadb7fc2fe909c352901ea24ae7bade45885aaa3a2991ef6 Nov 24 07:13:58 crc kubenswrapper[4809]: I1124 07:13:58.856328 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:13:58 crc kubenswrapper[4809]: I1124 07:13:58.902920 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb60935a-17c8-445c-b4ab-99786f623acb" path="/var/lib/kubelet/pods/fb60935a-17c8-445c-b4ab-99786f623acb/volumes" Nov 24 07:13:59 crc kubenswrapper[4809]: I1124 07:13:59.011750 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"111098ed-9d16-45e0-bb80-35f9383e327e","Type":"ContainerStarted","Data":"276bdc3c5eb76ef9fadb7fc2fe909c352901ea24ae7bade45885aaa3a2991ef6"} Nov 24 07:13:59 crc kubenswrapper[4809]: I1124 07:13:59.353725 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 24 07:13:59 crc kubenswrapper[4809]: I1124 07:13:59.441469 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 07:13:59 crc kubenswrapper[4809]: I1124 07:13:59.441536 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 07:14:00 crc kubenswrapper[4809]: I1124 07:14:00.020646 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"111098ed-9d16-45e0-bb80-35f9383e327e","Type":"ContainerStarted","Data":"ffb9df9f2296363c66395fba7c7aa3efbc5800406624655d76c1096454276be2"} Nov 24 07:14:00 crc kubenswrapper[4809]: I1124 07:14:00.043621 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.043605138 podStartE2EDuration="2.043605138s" podCreationTimestamp="2025-11-24 07:13:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:14:00.04331453 +0000 UTC m=+1139.943906255" watchObservedRunningTime="2025-11-24 07:14:00.043605138 +0000 UTC m=+1139.944196843" Nov 24 07:14:01 crc kubenswrapper[4809]: I1124 07:14:01.399420 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 07:14:01 crc kubenswrapper[4809]: I1124 07:14:01.399818 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="652f6c1f-9038-4993-a8c2-fb3e9a6a822f" containerName="kube-state-metrics" containerID="cri-o://464ead71dc25f530e7baac07291d1efa1b3c50cbfa75f7230f141ec025cefb49" gracePeriod=30 Nov 24 07:14:01 crc kubenswrapper[4809]: I1124 07:14:01.868231 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 07:14:01 crc kubenswrapper[4809]: I1124 07:14:01.964989 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d27cj\" (UniqueName: \"kubernetes.io/projected/652f6c1f-9038-4993-a8c2-fb3e9a6a822f-kube-api-access-d27cj\") pod \"652f6c1f-9038-4993-a8c2-fb3e9a6a822f\" (UID: \"652f6c1f-9038-4993-a8c2-fb3e9a6a822f\") " Nov 24 07:14:01 crc kubenswrapper[4809]: I1124 07:14:01.972283 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/652f6c1f-9038-4993-a8c2-fb3e9a6a822f-kube-api-access-d27cj" (OuterVolumeSpecName: "kube-api-access-d27cj") pod "652f6c1f-9038-4993-a8c2-fb3e9a6a822f" (UID: "652f6c1f-9038-4993-a8c2-fb3e9a6a822f"). InnerVolumeSpecName "kube-api-access-d27cj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.037700 4809 generic.go:334] "Generic (PLEG): container finished" podID="652f6c1f-9038-4993-a8c2-fb3e9a6a822f" containerID="464ead71dc25f530e7baac07291d1efa1b3c50cbfa75f7230f141ec025cefb49" exitCode=2 Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.037738 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"652f6c1f-9038-4993-a8c2-fb3e9a6a822f","Type":"ContainerDied","Data":"464ead71dc25f530e7baac07291d1efa1b3c50cbfa75f7230f141ec025cefb49"} Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.037761 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"652f6c1f-9038-4993-a8c2-fb3e9a6a822f","Type":"ContainerDied","Data":"e2f10f5bdc588121fc75edc1afa0205cbaa3121f567678594da10a3eca0b0335"} Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.037776 4809 scope.go:117] "RemoveContainer" containerID="464ead71dc25f530e7baac07291d1efa1b3c50cbfa75f7230f141ec025cefb49" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.037863 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.067540 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d27cj\" (UniqueName: \"kubernetes.io/projected/652f6c1f-9038-4993-a8c2-fb3e9a6a822f-kube-api-access-d27cj\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.070713 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.071104 4809 scope.go:117] "RemoveContainer" containerID="464ead71dc25f530e7baac07291d1efa1b3c50cbfa75f7230f141ec025cefb49" Nov 24 07:14:02 crc kubenswrapper[4809]: E1124 07:14:02.071455 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"464ead71dc25f530e7baac07291d1efa1b3c50cbfa75f7230f141ec025cefb49\": container with ID starting with 464ead71dc25f530e7baac07291d1efa1b3c50cbfa75f7230f141ec025cefb49 not found: ID does not exist" containerID="464ead71dc25f530e7baac07291d1efa1b3c50cbfa75f7230f141ec025cefb49" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.071492 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"464ead71dc25f530e7baac07291d1efa1b3c50cbfa75f7230f141ec025cefb49"} err="failed to get container status \"464ead71dc25f530e7baac07291d1efa1b3c50cbfa75f7230f141ec025cefb49\": rpc error: code = NotFound desc = could not find container \"464ead71dc25f530e7baac07291d1efa1b3c50cbfa75f7230f141ec025cefb49\": container with ID starting with 464ead71dc25f530e7baac07291d1efa1b3c50cbfa75f7230f141ec025cefb49 not found: ID does not exist" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.081403 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.094847 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 07:14:02 crc kubenswrapper[4809]: E1124 07:14:02.095258 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="652f6c1f-9038-4993-a8c2-fb3e9a6a822f" containerName="kube-state-metrics" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.095276 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="652f6c1f-9038-4993-a8c2-fb3e9a6a822f" containerName="kube-state-metrics" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.095451 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="652f6c1f-9038-4993-a8c2-fb3e9a6a822f" containerName="kube-state-metrics" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.096049 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.098517 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.098995 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.103850 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.168719 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/6f65b9e2-204d-455f-b8e1-43ccec98b7f6-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"6f65b9e2-204d-455f-b8e1-43ccec98b7f6\") " pod="openstack/kube-state-metrics-0" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.168815 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f65b9e2-204d-455f-b8e1-43ccec98b7f6-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"6f65b9e2-204d-455f-b8e1-43ccec98b7f6\") " pod="openstack/kube-state-metrics-0" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.168856 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnrzg\" (UniqueName: \"kubernetes.io/projected/6f65b9e2-204d-455f-b8e1-43ccec98b7f6-kube-api-access-nnrzg\") pod \"kube-state-metrics-0\" (UID: \"6f65b9e2-204d-455f-b8e1-43ccec98b7f6\") " pod="openstack/kube-state-metrics-0" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.168884 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f65b9e2-204d-455f-b8e1-43ccec98b7f6-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"6f65b9e2-204d-455f-b8e1-43ccec98b7f6\") " pod="openstack/kube-state-metrics-0" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.270341 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/6f65b9e2-204d-455f-b8e1-43ccec98b7f6-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"6f65b9e2-204d-455f-b8e1-43ccec98b7f6\") " pod="openstack/kube-state-metrics-0" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.270417 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f65b9e2-204d-455f-b8e1-43ccec98b7f6-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"6f65b9e2-204d-455f-b8e1-43ccec98b7f6\") " pod="openstack/kube-state-metrics-0" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.270449 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnrzg\" (UniqueName: \"kubernetes.io/projected/6f65b9e2-204d-455f-b8e1-43ccec98b7f6-kube-api-access-nnrzg\") pod \"kube-state-metrics-0\" (UID: \"6f65b9e2-204d-455f-b8e1-43ccec98b7f6\") " pod="openstack/kube-state-metrics-0" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.270470 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f65b9e2-204d-455f-b8e1-43ccec98b7f6-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"6f65b9e2-204d-455f-b8e1-43ccec98b7f6\") " pod="openstack/kube-state-metrics-0" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.274948 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/6f65b9e2-204d-455f-b8e1-43ccec98b7f6-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"6f65b9e2-204d-455f-b8e1-43ccec98b7f6\") " pod="openstack/kube-state-metrics-0" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.275722 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f65b9e2-204d-455f-b8e1-43ccec98b7f6-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"6f65b9e2-204d-455f-b8e1-43ccec98b7f6\") " pod="openstack/kube-state-metrics-0" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.276263 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f65b9e2-204d-455f-b8e1-43ccec98b7f6-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"6f65b9e2-204d-455f-b8e1-43ccec98b7f6\") " pod="openstack/kube-state-metrics-0" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.286068 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnrzg\" (UniqueName: \"kubernetes.io/projected/6f65b9e2-204d-455f-b8e1-43ccec98b7f6-kube-api-access-nnrzg\") pod \"kube-state-metrics-0\" (UID: \"6f65b9e2-204d-455f-b8e1-43ccec98b7f6\") " pod="openstack/kube-state-metrics-0" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.416518 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.877537 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 07:14:02 crc kubenswrapper[4809]: I1124 07:14:02.903947 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="652f6c1f-9038-4993-a8c2-fb3e9a6a822f" path="/var/lib/kubelet/pods/652f6c1f-9038-4993-a8c2-fb3e9a6a822f/volumes" Nov 24 07:14:03 crc kubenswrapper[4809]: I1124 07:14:03.047698 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6f65b9e2-204d-455f-b8e1-43ccec98b7f6","Type":"ContainerStarted","Data":"697fea19a97a552a52ae76c4d560dee77f35ae7143c4868eb58ba45563f4a690"} Nov 24 07:14:03 crc kubenswrapper[4809]: I1124 07:14:03.091072 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:14:03 crc kubenswrapper[4809]: I1124 07:14:03.091335 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" containerName="ceilometer-central-agent" containerID="cri-o://93e67478c477d5ae12fc6cc99f0c086e265f13bd9e6f24b78121eef62cfe436d" gracePeriod=30 Nov 24 07:14:03 crc kubenswrapper[4809]: I1124 07:14:03.091460 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" containerName="sg-core" containerID="cri-o://c6067d0e403b412b6a7cc202c5711bae48387b5dc487a7ffd0329dcff4961e1c" gracePeriod=30 Nov 24 07:14:03 crc kubenswrapper[4809]: I1124 07:14:03.091607 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" containerName="proxy-httpd" containerID="cri-o://b5b040ad8b4bc04437ba64141bdc8a0c315c0b4d8dbd58823f00a8966850e717" gracePeriod=30 Nov 24 07:14:03 crc kubenswrapper[4809]: I1124 07:14:03.091480 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" containerName="ceilometer-notification-agent" containerID="cri-o://f1eeac8352d8ca6d2059eaf6fa90f5fb9118557c95a77f09372f6772009b26d6" gracePeriod=30 Nov 24 07:14:03 crc kubenswrapper[4809]: I1124 07:14:03.420049 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 07:14:04 crc kubenswrapper[4809]: I1124 07:14:04.057547 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6f65b9e2-204d-455f-b8e1-43ccec98b7f6","Type":"ContainerStarted","Data":"402701ddfc6e4f304253544bf95ccd6a98b96f39914010b114e56b09477c9a5f"} Nov 24 07:14:04 crc kubenswrapper[4809]: I1124 07:14:04.057900 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 07:14:04 crc kubenswrapper[4809]: I1124 07:14:04.065143 4809 generic.go:334] "Generic (PLEG): container finished" podID="4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" containerID="b5b040ad8b4bc04437ba64141bdc8a0c315c0b4d8dbd58823f00a8966850e717" exitCode=0 Nov 24 07:14:04 crc kubenswrapper[4809]: I1124 07:14:04.065180 4809 generic.go:334] "Generic (PLEG): container finished" podID="4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" containerID="c6067d0e403b412b6a7cc202c5711bae48387b5dc487a7ffd0329dcff4961e1c" exitCode=2 Nov 24 07:14:04 crc kubenswrapper[4809]: I1124 07:14:04.065189 4809 generic.go:334] "Generic (PLEG): container finished" podID="4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" containerID="93e67478c477d5ae12fc6cc99f0c086e265f13bd9e6f24b78121eef62cfe436d" exitCode=0 Nov 24 07:14:04 crc kubenswrapper[4809]: I1124 07:14:04.065212 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12","Type":"ContainerDied","Data":"b5b040ad8b4bc04437ba64141bdc8a0c315c0b4d8dbd58823f00a8966850e717"} Nov 24 07:14:04 crc kubenswrapper[4809]: I1124 07:14:04.065240 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12","Type":"ContainerDied","Data":"c6067d0e403b412b6a7cc202c5711bae48387b5dc487a7ffd0329dcff4961e1c"} Nov 24 07:14:04 crc kubenswrapper[4809]: I1124 07:14:04.065254 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12","Type":"ContainerDied","Data":"93e67478c477d5ae12fc6cc99f0c086e265f13bd9e6f24b78121eef62cfe436d"} Nov 24 07:14:04 crc kubenswrapper[4809]: I1124 07:14:04.083859 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.680394247 podStartE2EDuration="2.083840273s" podCreationTimestamp="2025-11-24 07:14:02 +0000 UTC" firstStartedPulling="2025-11-24 07:14:02.902058032 +0000 UTC m=+1142.802649737" lastFinishedPulling="2025-11-24 07:14:03.305504058 +0000 UTC m=+1143.206095763" observedRunningTime="2025-11-24 07:14:04.075724558 +0000 UTC m=+1143.976316263" watchObservedRunningTime="2025-11-24 07:14:04.083840273 +0000 UTC m=+1143.984431988" Nov 24 07:14:04 crc kubenswrapper[4809]: I1124 07:14:04.418506 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 07:14:04 crc kubenswrapper[4809]: I1124 07:14:04.418855 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 07:14:04 crc kubenswrapper[4809]: I1124 07:14:04.441596 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 07:14:04 crc kubenswrapper[4809]: I1124 07:14:04.441643 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.527150 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2c0203aa-9e75-4bf4-8660-68f27b1b794b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.527149 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1286d658-6187-4509-9e24-1e0d6fbcab1e" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.527188 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2c0203aa-9e75-4bf4-8660-68f27b1b794b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.527225 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1286d658-6187-4509-9e24-1e0d6fbcab1e" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.772675 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.845183 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-run-httpd\") pod \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.845341 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-config-data\") pod \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.845379 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-scripts\") pod \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.845422 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llw4x\" (UniqueName: \"kubernetes.io/projected/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-kube-api-access-llw4x\") pod \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.845471 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-combined-ca-bundle\") pod \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.845576 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-log-httpd\") pod \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.845677 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-sg-core-conf-yaml\") pod \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\" (UID: \"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12\") " Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.847477 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" (UID: "4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.848158 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" (UID: "4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.852510 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-scripts" (OuterVolumeSpecName: "scripts") pod "4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" (UID: "4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.875182 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-kube-api-access-llw4x" (OuterVolumeSpecName: "kube-api-access-llw4x") pod "4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" (UID: "4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12"). InnerVolumeSpecName "kube-api-access-llw4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.902765 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" (UID: "4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.948091 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.948125 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llw4x\" (UniqueName: \"kubernetes.io/projected/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-kube-api-access-llw4x\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.948137 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.948148 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.948161 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.971985 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" (UID: "4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:05 crc kubenswrapper[4809]: I1124 07:14:05.997510 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-config-data" (OuterVolumeSpecName: "config-data") pod "4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" (UID: "4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.049912 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.049955 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.091401 4809 generic.go:334] "Generic (PLEG): container finished" podID="4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" containerID="f1eeac8352d8ca6d2059eaf6fa90f5fb9118557c95a77f09372f6772009b26d6" exitCode=0 Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.091452 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12","Type":"ContainerDied","Data":"f1eeac8352d8ca6d2059eaf6fa90f5fb9118557c95a77f09372f6772009b26d6"} Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.091462 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.091486 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12","Type":"ContainerDied","Data":"19539bd948969f6ad4f75c9beef8a27054770fe8039f316a8335bc5f1fa6fc1d"} Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.091511 4809 scope.go:117] "RemoveContainer" containerID="b5b040ad8b4bc04437ba64141bdc8a0c315c0b4d8dbd58823f00a8966850e717" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.118909 4809 scope.go:117] "RemoveContainer" containerID="c6067d0e403b412b6a7cc202c5711bae48387b5dc487a7ffd0329dcff4961e1c" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.130219 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.138880 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.148756 4809 scope.go:117] "RemoveContainer" containerID="f1eeac8352d8ca6d2059eaf6fa90f5fb9118557c95a77f09372f6772009b26d6" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.168250 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:14:06 crc kubenswrapper[4809]: E1124 07:14:06.168931 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" containerName="ceilometer-notification-agent" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.168958 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" containerName="ceilometer-notification-agent" Nov 24 07:14:06 crc kubenswrapper[4809]: E1124 07:14:06.169016 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" containerName="proxy-httpd" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.169023 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" containerName="proxy-httpd" Nov 24 07:14:06 crc kubenswrapper[4809]: E1124 07:14:06.169034 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" containerName="sg-core" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.169040 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" containerName="sg-core" Nov 24 07:14:06 crc kubenswrapper[4809]: E1124 07:14:06.169055 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" containerName="ceilometer-central-agent" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.169061 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" containerName="ceilometer-central-agent" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.169373 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" containerName="ceilometer-notification-agent" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.169392 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" containerName="proxy-httpd" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.169404 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" containerName="sg-core" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.169414 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" containerName="ceilometer-central-agent" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.176289 4809 scope.go:117] "RemoveContainer" containerID="93e67478c477d5ae12fc6cc99f0c086e265f13bd9e6f24b78121eef62cfe436d" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.183607 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.185524 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.186864 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.187454 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.253670 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-scripts\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.253832 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.253995 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-config-data\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.254030 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-run-httpd\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.254094 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-log-httpd\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.254164 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fhpb\" (UniqueName: \"kubernetes.io/projected/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-kube-api-access-7fhpb\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.254234 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.254309 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.269142 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.270423 4809 scope.go:117] "RemoveContainer" containerID="b5b040ad8b4bc04437ba64141bdc8a0c315c0b4d8dbd58823f00a8966850e717" Nov 24 07:14:06 crc kubenswrapper[4809]: E1124 07:14:06.270868 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5b040ad8b4bc04437ba64141bdc8a0c315c0b4d8dbd58823f00a8966850e717\": container with ID starting with b5b040ad8b4bc04437ba64141bdc8a0c315c0b4d8dbd58823f00a8966850e717 not found: ID does not exist" containerID="b5b040ad8b4bc04437ba64141bdc8a0c315c0b4d8dbd58823f00a8966850e717" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.270904 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5b040ad8b4bc04437ba64141bdc8a0c315c0b4d8dbd58823f00a8966850e717"} err="failed to get container status \"b5b040ad8b4bc04437ba64141bdc8a0c315c0b4d8dbd58823f00a8966850e717\": rpc error: code = NotFound desc = could not find container \"b5b040ad8b4bc04437ba64141bdc8a0c315c0b4d8dbd58823f00a8966850e717\": container with ID starting with b5b040ad8b4bc04437ba64141bdc8a0c315c0b4d8dbd58823f00a8966850e717 not found: ID does not exist" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.270951 4809 scope.go:117] "RemoveContainer" containerID="c6067d0e403b412b6a7cc202c5711bae48387b5dc487a7ffd0329dcff4961e1c" Nov 24 07:14:06 crc kubenswrapper[4809]: E1124 07:14:06.271333 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6067d0e403b412b6a7cc202c5711bae48387b5dc487a7ffd0329dcff4961e1c\": container with ID starting with c6067d0e403b412b6a7cc202c5711bae48387b5dc487a7ffd0329dcff4961e1c not found: ID does not exist" containerID="c6067d0e403b412b6a7cc202c5711bae48387b5dc487a7ffd0329dcff4961e1c" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.271360 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6067d0e403b412b6a7cc202c5711bae48387b5dc487a7ffd0329dcff4961e1c"} err="failed to get container status \"c6067d0e403b412b6a7cc202c5711bae48387b5dc487a7ffd0329dcff4961e1c\": rpc error: code = NotFound desc = could not find container \"c6067d0e403b412b6a7cc202c5711bae48387b5dc487a7ffd0329dcff4961e1c\": container with ID starting with c6067d0e403b412b6a7cc202c5711bae48387b5dc487a7ffd0329dcff4961e1c not found: ID does not exist" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.271400 4809 scope.go:117] "RemoveContainer" containerID="f1eeac8352d8ca6d2059eaf6fa90f5fb9118557c95a77f09372f6772009b26d6" Nov 24 07:14:06 crc kubenswrapper[4809]: E1124 07:14:06.271672 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1eeac8352d8ca6d2059eaf6fa90f5fb9118557c95a77f09372f6772009b26d6\": container with ID starting with f1eeac8352d8ca6d2059eaf6fa90f5fb9118557c95a77f09372f6772009b26d6 not found: ID does not exist" containerID="f1eeac8352d8ca6d2059eaf6fa90f5fb9118557c95a77f09372f6772009b26d6" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.271720 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1eeac8352d8ca6d2059eaf6fa90f5fb9118557c95a77f09372f6772009b26d6"} err="failed to get container status \"f1eeac8352d8ca6d2059eaf6fa90f5fb9118557c95a77f09372f6772009b26d6\": rpc error: code = NotFound desc = could not find container \"f1eeac8352d8ca6d2059eaf6fa90f5fb9118557c95a77f09372f6772009b26d6\": container with ID starting with f1eeac8352d8ca6d2059eaf6fa90f5fb9118557c95a77f09372f6772009b26d6 not found: ID does not exist" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.271747 4809 scope.go:117] "RemoveContainer" containerID="93e67478c477d5ae12fc6cc99f0c086e265f13bd9e6f24b78121eef62cfe436d" Nov 24 07:14:06 crc kubenswrapper[4809]: E1124 07:14:06.273028 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93e67478c477d5ae12fc6cc99f0c086e265f13bd9e6f24b78121eef62cfe436d\": container with ID starting with 93e67478c477d5ae12fc6cc99f0c086e265f13bd9e6f24b78121eef62cfe436d not found: ID does not exist" containerID="93e67478c477d5ae12fc6cc99f0c086e265f13bd9e6f24b78121eef62cfe436d" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.273060 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93e67478c477d5ae12fc6cc99f0c086e265f13bd9e6f24b78121eef62cfe436d"} err="failed to get container status \"93e67478c477d5ae12fc6cc99f0c086e265f13bd9e6f24b78121eef62cfe436d\": rpc error: code = NotFound desc = could not find container \"93e67478c477d5ae12fc6cc99f0c086e265f13bd9e6f24b78121eef62cfe436d\": container with ID starting with 93e67478c477d5ae12fc6cc99f0c086e265f13bd9e6f24b78121eef62cfe436d not found: ID does not exist" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.356873 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.356989 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.357110 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-scripts\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.357188 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.357356 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-config-data\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.357386 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-run-httpd\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.357456 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-log-httpd\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.357527 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fhpb\" (UniqueName: \"kubernetes.io/projected/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-kube-api-access-7fhpb\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.357982 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-run-httpd\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.358059 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-log-httpd\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.362796 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.363409 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-config-data\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.364687 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-scripts\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.365310 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.369527 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.377093 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fhpb\" (UniqueName: \"kubernetes.io/projected/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-kube-api-access-7fhpb\") pod \"ceilometer-0\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.541233 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:14:06 crc kubenswrapper[4809]: I1124 07:14:06.916016 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12" path="/var/lib/kubelet/pods/4aa2fc56-9abd-4a5a-ac78-8ad5ccec8e12/volumes" Nov 24 07:14:07 crc kubenswrapper[4809]: I1124 07:14:07.033085 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:14:07 crc kubenswrapper[4809]: I1124 07:14:07.102517 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2","Type":"ContainerStarted","Data":"66c415a383c8c66ba9a8865458983228e7614ad49c205cb399d949be62f132d2"} Nov 24 07:14:08 crc kubenswrapper[4809]: I1124 07:14:08.130632 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2","Type":"ContainerStarted","Data":"580e0450eda2c78d64e18d64e522a1608e489726b46db57dd5fd352b6df3afda"} Nov 24 07:14:08 crc kubenswrapper[4809]: I1124 07:14:08.419122 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 07:14:08 crc kubenswrapper[4809]: I1124 07:14:08.448584 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 07:14:09 crc kubenswrapper[4809]: I1124 07:14:09.141731 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2","Type":"ContainerStarted","Data":"a4669b3180e6fcd5d464baea9f3dd3493030fabb49fedacd16e46a4fd505ed98"} Nov 24 07:14:09 crc kubenswrapper[4809]: I1124 07:14:09.142056 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2","Type":"ContainerStarted","Data":"10a8f586ccca5dc3b00a839b30022b9f30601ca24d8b79fd6044e0f47d8a3bdc"} Nov 24 07:14:09 crc kubenswrapper[4809]: I1124 07:14:09.169994 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 07:14:11 crc kubenswrapper[4809]: I1124 07:14:11.176894 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2","Type":"ContainerStarted","Data":"aadeb5fa610e0d5d30eae59465719a6bac3bea1f9f12ead093f41e2c38108a0b"} Nov 24 07:14:11 crc kubenswrapper[4809]: I1124 07:14:11.178851 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 07:14:11 crc kubenswrapper[4809]: I1124 07:14:11.198261 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.168141035 podStartE2EDuration="5.198246479s" podCreationTimestamp="2025-11-24 07:14:06 +0000 UTC" firstStartedPulling="2025-11-24 07:14:07.041339341 +0000 UTC m=+1146.941931056" lastFinishedPulling="2025-11-24 07:14:10.071444795 +0000 UTC m=+1149.972036500" observedRunningTime="2025-11-24 07:14:11.197092228 +0000 UTC m=+1151.097683933" watchObservedRunningTime="2025-11-24 07:14:11.198246479 +0000 UTC m=+1151.098838184" Nov 24 07:14:12 crc kubenswrapper[4809]: I1124 07:14:12.431439 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 07:14:14 crc kubenswrapper[4809]: I1124 07:14:14.421300 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 07:14:14 crc kubenswrapper[4809]: I1124 07:14:14.422394 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 07:14:14 crc kubenswrapper[4809]: I1124 07:14:14.425372 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 07:14:14 crc kubenswrapper[4809]: I1124 07:14:14.426426 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 07:14:14 crc kubenswrapper[4809]: I1124 07:14:14.452184 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 07:14:14 crc kubenswrapper[4809]: I1124 07:14:14.452294 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 07:14:14 crc kubenswrapper[4809]: I1124 07:14:14.462105 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 07:14:14 crc kubenswrapper[4809]: I1124 07:14:14.466832 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.211596 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.215318 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.405618 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-w774p"] Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.407311 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.423199 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-w774p"] Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.429903 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-w774p\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.429980 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-w774p\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.430045 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-config\") pod \"dnsmasq-dns-89c5cd4d5-w774p\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.430097 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwmf4\" (UniqueName: \"kubernetes.io/projected/e535ca4c-f961-44f0-a031-61d4264a984c-kube-api-access-cwmf4\") pod \"dnsmasq-dns-89c5cd4d5-w774p\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.430120 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-w774p\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.430158 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-w774p\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.531602 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-config\") pod \"dnsmasq-dns-89c5cd4d5-w774p\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.532065 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwmf4\" (UniqueName: \"kubernetes.io/projected/e535ca4c-f961-44f0-a031-61d4264a984c-kube-api-access-cwmf4\") pod \"dnsmasq-dns-89c5cd4d5-w774p\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.532093 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-w774p\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.532138 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-w774p\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.532183 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-w774p\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.532223 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-w774p\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.533303 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-w774p\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.533322 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-w774p\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.533388 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-w774p\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.533556 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-w774p\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.533887 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-config\") pod \"dnsmasq-dns-89c5cd4d5-w774p\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.553709 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwmf4\" (UniqueName: \"kubernetes.io/projected/e535ca4c-f961-44f0-a031-61d4264a984c-kube-api-access-cwmf4\") pod \"dnsmasq-dns-89c5cd4d5-w774p\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:14:15 crc kubenswrapper[4809]: I1124 07:14:15.745651 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:14:16 crc kubenswrapper[4809]: I1124 07:14:16.246006 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-w774p"] Nov 24 07:14:16 crc kubenswrapper[4809]: W1124 07:14:16.250725 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode535ca4c_f961_44f0_a031_61d4264a984c.slice/crio-66c03ee2d9f6909b36e505a5fc232e3106e1493ea382fa35ee00f5f33e959163 WatchSource:0}: Error finding container 66c03ee2d9f6909b36e505a5fc232e3106e1493ea382fa35ee00f5f33e959163: Status 404 returned error can't find the container with id 66c03ee2d9f6909b36e505a5fc232e3106e1493ea382fa35ee00f5f33e959163 Nov 24 07:14:17 crc kubenswrapper[4809]: I1124 07:14:17.245005 4809 generic.go:334] "Generic (PLEG): container finished" podID="e535ca4c-f961-44f0-a031-61d4264a984c" containerID="4b795101827cd4186f9dda1abb76897656daeedd6ece3f3f35b3eb0f465b0638" exitCode=0 Nov 24 07:14:17 crc kubenswrapper[4809]: I1124 07:14:17.245424 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" event={"ID":"e535ca4c-f961-44f0-a031-61d4264a984c","Type":"ContainerDied","Data":"4b795101827cd4186f9dda1abb76897656daeedd6ece3f3f35b3eb0f465b0638"} Nov 24 07:14:17 crc kubenswrapper[4809]: I1124 07:14:17.246810 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" event={"ID":"e535ca4c-f961-44f0-a031-61d4264a984c","Type":"ContainerStarted","Data":"66c03ee2d9f6909b36e505a5fc232e3106e1493ea382fa35ee00f5f33e959163"} Nov 24 07:14:17 crc kubenswrapper[4809]: I1124 07:14:17.264388 4809 generic.go:334] "Generic (PLEG): container finished" podID="583f5bd5-82d6-4d48-819a-3f84049011c9" containerID="4cd3d9f60485af0ee52d3e65e02fda14955ef8130cccd9b4978f7329249cf359" exitCode=137 Nov 24 07:14:17 crc kubenswrapper[4809]: I1124 07:14:17.264630 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"583f5bd5-82d6-4d48-819a-3f84049011c9","Type":"ContainerDied","Data":"4cd3d9f60485af0ee52d3e65e02fda14955ef8130cccd9b4978f7329249cf359"} Nov 24 07:14:17 crc kubenswrapper[4809]: I1124 07:14:17.264684 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"583f5bd5-82d6-4d48-819a-3f84049011c9","Type":"ContainerDied","Data":"0efeda542d0bd46d9b4533002aa182edd248c7b9b856b9e738ef8d01c26d8405"} Nov 24 07:14:17 crc kubenswrapper[4809]: I1124 07:14:17.264695 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0efeda542d0bd46d9b4533002aa182edd248c7b9b856b9e738ef8d01c26d8405" Nov 24 07:14:17 crc kubenswrapper[4809]: I1124 07:14:17.363447 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:14:17 crc kubenswrapper[4809]: I1124 07:14:17.478316 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/583f5bd5-82d6-4d48-819a-3f84049011c9-config-data\") pod \"583f5bd5-82d6-4d48-819a-3f84049011c9\" (UID: \"583f5bd5-82d6-4d48-819a-3f84049011c9\") " Nov 24 07:14:17 crc kubenswrapper[4809]: I1124 07:14:17.478412 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/583f5bd5-82d6-4d48-819a-3f84049011c9-combined-ca-bundle\") pod \"583f5bd5-82d6-4d48-819a-3f84049011c9\" (UID: \"583f5bd5-82d6-4d48-819a-3f84049011c9\") " Nov 24 07:14:17 crc kubenswrapper[4809]: I1124 07:14:17.478625 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5b2wn\" (UniqueName: \"kubernetes.io/projected/583f5bd5-82d6-4d48-819a-3f84049011c9-kube-api-access-5b2wn\") pod \"583f5bd5-82d6-4d48-819a-3f84049011c9\" (UID: \"583f5bd5-82d6-4d48-819a-3f84049011c9\") " Nov 24 07:14:17 crc kubenswrapper[4809]: I1124 07:14:17.483635 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/583f5bd5-82d6-4d48-819a-3f84049011c9-kube-api-access-5b2wn" (OuterVolumeSpecName: "kube-api-access-5b2wn") pod "583f5bd5-82d6-4d48-819a-3f84049011c9" (UID: "583f5bd5-82d6-4d48-819a-3f84049011c9"). InnerVolumeSpecName "kube-api-access-5b2wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:14:17 crc kubenswrapper[4809]: I1124 07:14:17.508468 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/583f5bd5-82d6-4d48-819a-3f84049011c9-config-data" (OuterVolumeSpecName: "config-data") pod "583f5bd5-82d6-4d48-819a-3f84049011c9" (UID: "583f5bd5-82d6-4d48-819a-3f84049011c9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:17 crc kubenswrapper[4809]: I1124 07:14:17.508562 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/583f5bd5-82d6-4d48-819a-3f84049011c9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "583f5bd5-82d6-4d48-819a-3f84049011c9" (UID: "583f5bd5-82d6-4d48-819a-3f84049011c9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:17 crc kubenswrapper[4809]: I1124 07:14:17.580953 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5b2wn\" (UniqueName: \"kubernetes.io/projected/583f5bd5-82d6-4d48-819a-3f84049011c9-kube-api-access-5b2wn\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:17 crc kubenswrapper[4809]: I1124 07:14:17.581258 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/583f5bd5-82d6-4d48-819a-3f84049011c9-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:17 crc kubenswrapper[4809]: I1124 07:14:17.581270 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/583f5bd5-82d6-4d48-819a-3f84049011c9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:17 crc kubenswrapper[4809]: I1124 07:14:17.769112 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:14:17 crc kubenswrapper[4809]: I1124 07:14:17.792448 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:14:17 crc kubenswrapper[4809]: I1124 07:14:17.792710 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" containerName="ceilometer-central-agent" containerID="cri-o://580e0450eda2c78d64e18d64e522a1608e489726b46db57dd5fd352b6df3afda" gracePeriod=30 Nov 24 07:14:17 crc kubenswrapper[4809]: I1124 07:14:17.793114 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" containerName="proxy-httpd" containerID="cri-o://aadeb5fa610e0d5d30eae59465719a6bac3bea1f9f12ead093f41e2c38108a0b" gracePeriod=30 Nov 24 07:14:17 crc kubenswrapper[4809]: I1124 07:14:17.793184 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" containerName="sg-core" containerID="cri-o://a4669b3180e6fcd5d464baea9f3dd3493030fabb49fedacd16e46a4fd505ed98" gracePeriod=30 Nov 24 07:14:17 crc kubenswrapper[4809]: I1124 07:14:17.793218 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" containerName="ceilometer-notification-agent" containerID="cri-o://10a8f586ccca5dc3b00a839b30022b9f30601ca24d8b79fd6044e0f47d8a3bdc" gracePeriod=30 Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.276472 4809 generic.go:334] "Generic (PLEG): container finished" podID="fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" containerID="aadeb5fa610e0d5d30eae59465719a6bac3bea1f9f12ead093f41e2c38108a0b" exitCode=0 Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.276519 4809 generic.go:334] "Generic (PLEG): container finished" podID="fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" containerID="a4669b3180e6fcd5d464baea9f3dd3493030fabb49fedacd16e46a4fd505ed98" exitCode=2 Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.276528 4809 generic.go:334] "Generic (PLEG): container finished" podID="fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" containerID="580e0450eda2c78d64e18d64e522a1608e489726b46db57dd5fd352b6df3afda" exitCode=0 Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.276551 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2","Type":"ContainerDied","Data":"aadeb5fa610e0d5d30eae59465719a6bac3bea1f9f12ead093f41e2c38108a0b"} Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.276600 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2","Type":"ContainerDied","Data":"a4669b3180e6fcd5d464baea9f3dd3493030fabb49fedacd16e46a4fd505ed98"} Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.276615 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2","Type":"ContainerDied","Data":"580e0450eda2c78d64e18d64e522a1608e489726b46db57dd5fd352b6df3afda"} Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.279901 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" event={"ID":"e535ca4c-f961-44f0-a031-61d4264a984c","Type":"ContainerStarted","Data":"f9ff0326dfa3960ea21b7f760328dcba59a13ea8bfdf0dcdd7c217a9570755e0"} Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.279958 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.282371 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2c0203aa-9e75-4bf4-8660-68f27b1b794b" containerName="nova-api-api" containerID="cri-o://a8c9a69b328f2e57ec6455cd0c4ee70ec9c896be31e4103c6cac1edb7d7381ca" gracePeriod=30 Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.282594 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2c0203aa-9e75-4bf4-8660-68f27b1b794b" containerName="nova-api-log" containerID="cri-o://19bf512267a342fe3d48842756d71bdb73d4717364dfec75d2e3e6d89ef9578f" gracePeriod=30 Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.315288 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" podStartSLOduration=3.315268075 podStartE2EDuration="3.315268075s" podCreationTimestamp="2025-11-24 07:14:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:14:18.310102078 +0000 UTC m=+1158.210693793" watchObservedRunningTime="2025-11-24 07:14:18.315268075 +0000 UTC m=+1158.215859780" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.354352 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.366287 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.392288 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 07:14:18 crc kubenswrapper[4809]: E1124 07:14:18.392705 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="583f5bd5-82d6-4d48-819a-3f84049011c9" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.392722 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="583f5bd5-82d6-4d48-819a-3f84049011c9" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.392922 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="583f5bd5-82d6-4d48-819a-3f84049011c9" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.393557 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.395499 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.400529 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.400895 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.401152 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.504373 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0f2a430-fd7f-419b-a07c-283144e6f0f2-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0f2a430-fd7f-419b-a07c-283144e6f0f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.504472 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7px8\" (UniqueName: \"kubernetes.io/projected/c0f2a430-fd7f-419b-a07c-283144e6f0f2-kube-api-access-t7px8\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0f2a430-fd7f-419b-a07c-283144e6f0f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.504679 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0f2a430-fd7f-419b-a07c-283144e6f0f2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0f2a430-fd7f-419b-a07c-283144e6f0f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.504748 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0f2a430-fd7f-419b-a07c-283144e6f0f2-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0f2a430-fd7f-419b-a07c-283144e6f0f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.504947 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0f2a430-fd7f-419b-a07c-283144e6f0f2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0f2a430-fd7f-419b-a07c-283144e6f0f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.607859 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0f2a430-fd7f-419b-a07c-283144e6f0f2-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0f2a430-fd7f-419b-a07c-283144e6f0f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.608209 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7px8\" (UniqueName: \"kubernetes.io/projected/c0f2a430-fd7f-419b-a07c-283144e6f0f2-kube-api-access-t7px8\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0f2a430-fd7f-419b-a07c-283144e6f0f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.608310 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0f2a430-fd7f-419b-a07c-283144e6f0f2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0f2a430-fd7f-419b-a07c-283144e6f0f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.608339 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0f2a430-fd7f-419b-a07c-283144e6f0f2-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0f2a430-fd7f-419b-a07c-283144e6f0f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.609097 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0f2a430-fd7f-419b-a07c-283144e6f0f2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0f2a430-fd7f-419b-a07c-283144e6f0f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.615302 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0f2a430-fd7f-419b-a07c-283144e6f0f2-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0f2a430-fd7f-419b-a07c-283144e6f0f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.616115 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0f2a430-fd7f-419b-a07c-283144e6f0f2-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0f2a430-fd7f-419b-a07c-283144e6f0f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.619229 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0f2a430-fd7f-419b-a07c-283144e6f0f2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0f2a430-fd7f-419b-a07c-283144e6f0f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.619472 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0f2a430-fd7f-419b-a07c-283144e6f0f2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0f2a430-fd7f-419b-a07c-283144e6f0f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.635394 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7px8\" (UniqueName: \"kubernetes.io/projected/c0f2a430-fd7f-419b-a07c-283144e6f0f2-kube-api-access-t7px8\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0f2a430-fd7f-419b-a07c-283144e6f0f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.760219 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.874320 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.904934 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="583f5bd5-82d6-4d48-819a-3f84049011c9" path="/var/lib/kubelet/pods/583f5bd5-82d6-4d48-819a-3f84049011c9/volumes" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.913929 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fhpb\" (UniqueName: \"kubernetes.io/projected/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-kube-api-access-7fhpb\") pod \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.914017 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-ceilometer-tls-certs\") pod \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.914054 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-run-httpd\") pod \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.914087 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-combined-ca-bundle\") pod \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.914155 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-sg-core-conf-yaml\") pod \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.914244 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-scripts\") pod \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.914281 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-log-httpd\") pod \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.914352 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" (UID: "fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.914370 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-config-data\") pod \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\" (UID: \"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2\") " Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.914976 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" (UID: "fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.915200 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.915216 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.928990 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-kube-api-access-7fhpb" (OuterVolumeSpecName: "kube-api-access-7fhpb") pod "fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" (UID: "fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2"). InnerVolumeSpecName "kube-api-access-7fhpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:14:18 crc kubenswrapper[4809]: I1124 07:14:18.952323 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-scripts" (OuterVolumeSpecName: "scripts") pod "fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" (UID: "fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.012885 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" (UID: "fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.016742 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.016766 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fhpb\" (UniqueName: \"kubernetes.io/projected/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-kube-api-access-7fhpb\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.016776 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.031838 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" (UID: "fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.053755 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" (UID: "fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.064492 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-config-data" (OuterVolumeSpecName: "config-data") pod "fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" (UID: "fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.118395 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.118664 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.118728 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.225263 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 07:14:19 crc kubenswrapper[4809]: W1124 07:14:19.228231 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0f2a430_fd7f_419b_a07c_283144e6f0f2.slice/crio-844eea5292ae2fa0fe0be5a00b37e1e6721b5a06fd092d4b8057c49fadac3044 WatchSource:0}: Error finding container 844eea5292ae2fa0fe0be5a00b37e1e6721b5a06fd092d4b8057c49fadac3044: Status 404 returned error can't find the container with id 844eea5292ae2fa0fe0be5a00b37e1e6721b5a06fd092d4b8057c49fadac3044 Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.306664 4809 generic.go:334] "Generic (PLEG): container finished" podID="fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" containerID="10a8f586ccca5dc3b00a839b30022b9f30601ca24d8b79fd6044e0f47d8a3bdc" exitCode=0 Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.306784 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2","Type":"ContainerDied","Data":"10a8f586ccca5dc3b00a839b30022b9f30601ca24d8b79fd6044e0f47d8a3bdc"} Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.306804 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.306817 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2","Type":"ContainerDied","Data":"66c415a383c8c66ba9a8865458983228e7614ad49c205cb399d949be62f132d2"} Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.306861 4809 scope.go:117] "RemoveContainer" containerID="aadeb5fa610e0d5d30eae59465719a6bac3bea1f9f12ead093f41e2c38108a0b" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.309642 4809 generic.go:334] "Generic (PLEG): container finished" podID="2c0203aa-9e75-4bf4-8660-68f27b1b794b" containerID="19bf512267a342fe3d48842756d71bdb73d4717364dfec75d2e3e6d89ef9578f" exitCode=143 Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.309694 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2c0203aa-9e75-4bf4-8660-68f27b1b794b","Type":"ContainerDied","Data":"19bf512267a342fe3d48842756d71bdb73d4717364dfec75d2e3e6d89ef9578f"} Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.312657 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c0f2a430-fd7f-419b-a07c-283144e6f0f2","Type":"ContainerStarted","Data":"844eea5292ae2fa0fe0be5a00b37e1e6721b5a06fd092d4b8057c49fadac3044"} Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.312894 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.331402 4809 scope.go:117] "RemoveContainer" containerID="a4669b3180e6fcd5d464baea9f3dd3493030fabb49fedacd16e46a4fd505ed98" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.341355 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.357525 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.365353 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:14:19 crc kubenswrapper[4809]: E1124 07:14:19.365747 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" containerName="sg-core" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.366064 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" containerName="sg-core" Nov 24 07:14:19 crc kubenswrapper[4809]: E1124 07:14:19.366093 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" containerName="proxy-httpd" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.366099 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" containerName="proxy-httpd" Nov 24 07:14:19 crc kubenswrapper[4809]: E1124 07:14:19.366114 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" containerName="ceilometer-central-agent" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.366121 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" containerName="ceilometer-central-agent" Nov 24 07:14:19 crc kubenswrapper[4809]: E1124 07:14:19.366131 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" containerName="ceilometer-notification-agent" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.366137 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" containerName="ceilometer-notification-agent" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.366304 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" containerName="proxy-httpd" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.366312 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" containerName="ceilometer-notification-agent" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.366323 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" containerName="sg-core" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.366335 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" containerName="ceilometer-central-agent" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.369009 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.371988 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.372141 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.372205 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.375151 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.388792 4809 scope.go:117] "RemoveContainer" containerID="10a8f586ccca5dc3b00a839b30022b9f30601ca24d8b79fd6044e0f47d8a3bdc" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.414956 4809 scope.go:117] "RemoveContainer" containerID="580e0450eda2c78d64e18d64e522a1608e489726b46db57dd5fd352b6df3afda" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.423983 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-run-httpd\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.424041 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.424076 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-log-httpd\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.424124 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.424237 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-scripts\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.424299 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpx4w\" (UniqueName: \"kubernetes.io/projected/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-kube-api-access-bpx4w\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.424325 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.424356 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-config-data\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.445568 4809 scope.go:117] "RemoveContainer" containerID="aadeb5fa610e0d5d30eae59465719a6bac3bea1f9f12ead093f41e2c38108a0b" Nov 24 07:14:19 crc kubenswrapper[4809]: E1124 07:14:19.447052 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aadeb5fa610e0d5d30eae59465719a6bac3bea1f9f12ead093f41e2c38108a0b\": container with ID starting with aadeb5fa610e0d5d30eae59465719a6bac3bea1f9f12ead093f41e2c38108a0b not found: ID does not exist" containerID="aadeb5fa610e0d5d30eae59465719a6bac3bea1f9f12ead093f41e2c38108a0b" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.447084 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aadeb5fa610e0d5d30eae59465719a6bac3bea1f9f12ead093f41e2c38108a0b"} err="failed to get container status \"aadeb5fa610e0d5d30eae59465719a6bac3bea1f9f12ead093f41e2c38108a0b\": rpc error: code = NotFound desc = could not find container \"aadeb5fa610e0d5d30eae59465719a6bac3bea1f9f12ead093f41e2c38108a0b\": container with ID starting with aadeb5fa610e0d5d30eae59465719a6bac3bea1f9f12ead093f41e2c38108a0b not found: ID does not exist" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.447103 4809 scope.go:117] "RemoveContainer" containerID="a4669b3180e6fcd5d464baea9f3dd3493030fabb49fedacd16e46a4fd505ed98" Nov 24 07:14:19 crc kubenswrapper[4809]: E1124 07:14:19.447407 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4669b3180e6fcd5d464baea9f3dd3493030fabb49fedacd16e46a4fd505ed98\": container with ID starting with a4669b3180e6fcd5d464baea9f3dd3493030fabb49fedacd16e46a4fd505ed98 not found: ID does not exist" containerID="a4669b3180e6fcd5d464baea9f3dd3493030fabb49fedacd16e46a4fd505ed98" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.447426 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4669b3180e6fcd5d464baea9f3dd3493030fabb49fedacd16e46a4fd505ed98"} err="failed to get container status \"a4669b3180e6fcd5d464baea9f3dd3493030fabb49fedacd16e46a4fd505ed98\": rpc error: code = NotFound desc = could not find container \"a4669b3180e6fcd5d464baea9f3dd3493030fabb49fedacd16e46a4fd505ed98\": container with ID starting with a4669b3180e6fcd5d464baea9f3dd3493030fabb49fedacd16e46a4fd505ed98 not found: ID does not exist" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.447438 4809 scope.go:117] "RemoveContainer" containerID="10a8f586ccca5dc3b00a839b30022b9f30601ca24d8b79fd6044e0f47d8a3bdc" Nov 24 07:14:19 crc kubenswrapper[4809]: E1124 07:14:19.458360 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10a8f586ccca5dc3b00a839b30022b9f30601ca24d8b79fd6044e0f47d8a3bdc\": container with ID starting with 10a8f586ccca5dc3b00a839b30022b9f30601ca24d8b79fd6044e0f47d8a3bdc not found: ID does not exist" containerID="10a8f586ccca5dc3b00a839b30022b9f30601ca24d8b79fd6044e0f47d8a3bdc" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.458408 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10a8f586ccca5dc3b00a839b30022b9f30601ca24d8b79fd6044e0f47d8a3bdc"} err="failed to get container status \"10a8f586ccca5dc3b00a839b30022b9f30601ca24d8b79fd6044e0f47d8a3bdc\": rpc error: code = NotFound desc = could not find container \"10a8f586ccca5dc3b00a839b30022b9f30601ca24d8b79fd6044e0f47d8a3bdc\": container with ID starting with 10a8f586ccca5dc3b00a839b30022b9f30601ca24d8b79fd6044e0f47d8a3bdc not found: ID does not exist" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.458433 4809 scope.go:117] "RemoveContainer" containerID="580e0450eda2c78d64e18d64e522a1608e489726b46db57dd5fd352b6df3afda" Nov 24 07:14:19 crc kubenswrapper[4809]: E1124 07:14:19.458828 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"580e0450eda2c78d64e18d64e522a1608e489726b46db57dd5fd352b6df3afda\": container with ID starting with 580e0450eda2c78d64e18d64e522a1608e489726b46db57dd5fd352b6df3afda not found: ID does not exist" containerID="580e0450eda2c78d64e18d64e522a1608e489726b46db57dd5fd352b6df3afda" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.458853 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"580e0450eda2c78d64e18d64e522a1608e489726b46db57dd5fd352b6df3afda"} err="failed to get container status \"580e0450eda2c78d64e18d64e522a1608e489726b46db57dd5fd352b6df3afda\": rpc error: code = NotFound desc = could not find container \"580e0450eda2c78d64e18d64e522a1608e489726b46db57dd5fd352b6df3afda\": container with ID starting with 580e0450eda2c78d64e18d64e522a1608e489726b46db57dd5fd352b6df3afda not found: ID does not exist" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.525540 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-scripts\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.525605 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpx4w\" (UniqueName: \"kubernetes.io/projected/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-kube-api-access-bpx4w\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.525629 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.525654 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-config-data\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.525687 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-run-httpd\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.525713 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.525735 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-log-httpd\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.525768 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.526718 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-run-httpd\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.527004 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-log-httpd\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.530707 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.532913 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-scripts\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.533982 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-config-data\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.534118 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.534478 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.547699 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpx4w\" (UniqueName: \"kubernetes.io/projected/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-kube-api-access-bpx4w\") pod \"ceilometer-0\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " pod="openstack/ceilometer-0" Nov 24 07:14:19 crc kubenswrapper[4809]: I1124 07:14:19.685775 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:14:20 crc kubenswrapper[4809]: I1124 07:14:20.184754 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:14:20 crc kubenswrapper[4809]: I1124 07:14:20.207981 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:14:20 crc kubenswrapper[4809]: I1124 07:14:20.322055 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52","Type":"ContainerStarted","Data":"f23ab21a2611f62431ae3cceae0365a16a05e7bb981b9a13ba8c5a3520088ee9"} Nov 24 07:14:20 crc kubenswrapper[4809]: I1124 07:14:20.324631 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c0f2a430-fd7f-419b-a07c-283144e6f0f2","Type":"ContainerStarted","Data":"5e11da768201f975c7883afffb75522d4103d313ae71a1245d79e6c32dd5ec79"} Nov 24 07:14:20 crc kubenswrapper[4809]: I1124 07:14:20.351497 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.351476718 podStartE2EDuration="2.351476718s" podCreationTimestamp="2025-11-24 07:14:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:14:20.33950588 +0000 UTC m=+1160.240097595" watchObservedRunningTime="2025-11-24 07:14:20.351476718 +0000 UTC m=+1160.252068433" Nov 24 07:14:20 crc kubenswrapper[4809]: I1124 07:14:20.905580 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2" path="/var/lib/kubelet/pods/fac7cdfc-ed79-4fc2-aa3b-abbc842f0fe2/volumes" Nov 24 07:14:21 crc kubenswrapper[4809]: I1124 07:14:21.336529 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52","Type":"ContainerStarted","Data":"6f443b63c63f414412d4726e5018d4498ace22da416c81c9def3225bf2cbcbc3"} Nov 24 07:14:21 crc kubenswrapper[4809]: I1124 07:14:21.823198 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.003581 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c0203aa-9e75-4bf4-8660-68f27b1b794b-logs\") pod \"2c0203aa-9e75-4bf4-8660-68f27b1b794b\" (UID: \"2c0203aa-9e75-4bf4-8660-68f27b1b794b\") " Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.003880 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvgkr\" (UniqueName: \"kubernetes.io/projected/2c0203aa-9e75-4bf4-8660-68f27b1b794b-kube-api-access-zvgkr\") pod \"2c0203aa-9e75-4bf4-8660-68f27b1b794b\" (UID: \"2c0203aa-9e75-4bf4-8660-68f27b1b794b\") " Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.004030 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c0203aa-9e75-4bf4-8660-68f27b1b794b-combined-ca-bundle\") pod \"2c0203aa-9e75-4bf4-8660-68f27b1b794b\" (UID: \"2c0203aa-9e75-4bf4-8660-68f27b1b794b\") " Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.004169 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c0203aa-9e75-4bf4-8660-68f27b1b794b-config-data\") pod \"2c0203aa-9e75-4bf4-8660-68f27b1b794b\" (UID: \"2c0203aa-9e75-4bf4-8660-68f27b1b794b\") " Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.004195 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c0203aa-9e75-4bf4-8660-68f27b1b794b-logs" (OuterVolumeSpecName: "logs") pod "2c0203aa-9e75-4bf4-8660-68f27b1b794b" (UID: "2c0203aa-9e75-4bf4-8660-68f27b1b794b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.004760 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c0203aa-9e75-4bf4-8660-68f27b1b794b-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.037127 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c0203aa-9e75-4bf4-8660-68f27b1b794b-kube-api-access-zvgkr" (OuterVolumeSpecName: "kube-api-access-zvgkr") pod "2c0203aa-9e75-4bf4-8660-68f27b1b794b" (UID: "2c0203aa-9e75-4bf4-8660-68f27b1b794b"). InnerVolumeSpecName "kube-api-access-zvgkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.042985 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c0203aa-9e75-4bf4-8660-68f27b1b794b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c0203aa-9e75-4bf4-8660-68f27b1b794b" (UID: "2c0203aa-9e75-4bf4-8660-68f27b1b794b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.048207 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c0203aa-9e75-4bf4-8660-68f27b1b794b-config-data" (OuterVolumeSpecName: "config-data") pod "2c0203aa-9e75-4bf4-8660-68f27b1b794b" (UID: "2c0203aa-9e75-4bf4-8660-68f27b1b794b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.106320 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c0203aa-9e75-4bf4-8660-68f27b1b794b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.106540 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvgkr\" (UniqueName: \"kubernetes.io/projected/2c0203aa-9e75-4bf4-8660-68f27b1b794b-kube-api-access-zvgkr\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.106635 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c0203aa-9e75-4bf4-8660-68f27b1b794b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.347218 4809 generic.go:334] "Generic (PLEG): container finished" podID="2c0203aa-9e75-4bf4-8660-68f27b1b794b" containerID="a8c9a69b328f2e57ec6455cd0c4ee70ec9c896be31e4103c6cac1edb7d7381ca" exitCode=0 Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.347292 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.347305 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2c0203aa-9e75-4bf4-8660-68f27b1b794b","Type":"ContainerDied","Data":"a8c9a69b328f2e57ec6455cd0c4ee70ec9c896be31e4103c6cac1edb7d7381ca"} Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.348361 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2c0203aa-9e75-4bf4-8660-68f27b1b794b","Type":"ContainerDied","Data":"684bc7f2cbd1c89af75e2704aa5f285e26d138c82a16653bf1a4bf90b9cdf5bc"} Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.348397 4809 scope.go:117] "RemoveContainer" containerID="a8c9a69b328f2e57ec6455cd0c4ee70ec9c896be31e4103c6cac1edb7d7381ca" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.351859 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52","Type":"ContainerStarted","Data":"0e5e353e5533e9b43aa58a962905998869290a3459004a4dc8e03bf9f8ac619c"} Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.377418 4809 scope.go:117] "RemoveContainer" containerID="19bf512267a342fe3d48842756d71bdb73d4717364dfec75d2e3e6d89ef9578f" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.409098 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.439884 4809 scope.go:117] "RemoveContainer" containerID="a8c9a69b328f2e57ec6455cd0c4ee70ec9c896be31e4103c6cac1edb7d7381ca" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.441685 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:14:22 crc kubenswrapper[4809]: E1124 07:14:22.443479 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8c9a69b328f2e57ec6455cd0c4ee70ec9c896be31e4103c6cac1edb7d7381ca\": container with ID starting with a8c9a69b328f2e57ec6455cd0c4ee70ec9c896be31e4103c6cac1edb7d7381ca not found: ID does not exist" containerID="a8c9a69b328f2e57ec6455cd0c4ee70ec9c896be31e4103c6cac1edb7d7381ca" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.443527 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8c9a69b328f2e57ec6455cd0c4ee70ec9c896be31e4103c6cac1edb7d7381ca"} err="failed to get container status \"a8c9a69b328f2e57ec6455cd0c4ee70ec9c896be31e4103c6cac1edb7d7381ca\": rpc error: code = NotFound desc = could not find container \"a8c9a69b328f2e57ec6455cd0c4ee70ec9c896be31e4103c6cac1edb7d7381ca\": container with ID starting with a8c9a69b328f2e57ec6455cd0c4ee70ec9c896be31e4103c6cac1edb7d7381ca not found: ID does not exist" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.443554 4809 scope.go:117] "RemoveContainer" containerID="19bf512267a342fe3d48842756d71bdb73d4717364dfec75d2e3e6d89ef9578f" Nov 24 07:14:22 crc kubenswrapper[4809]: E1124 07:14:22.447654 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19bf512267a342fe3d48842756d71bdb73d4717364dfec75d2e3e6d89ef9578f\": container with ID starting with 19bf512267a342fe3d48842756d71bdb73d4717364dfec75d2e3e6d89ef9578f not found: ID does not exist" containerID="19bf512267a342fe3d48842756d71bdb73d4717364dfec75d2e3e6d89ef9578f" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.447709 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19bf512267a342fe3d48842756d71bdb73d4717364dfec75d2e3e6d89ef9578f"} err="failed to get container status \"19bf512267a342fe3d48842756d71bdb73d4717364dfec75d2e3e6d89ef9578f\": rpc error: code = NotFound desc = could not find container \"19bf512267a342fe3d48842756d71bdb73d4717364dfec75d2e3e6d89ef9578f\": container with ID starting with 19bf512267a342fe3d48842756d71bdb73d4717364dfec75d2e3e6d89ef9578f not found: ID does not exist" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.456848 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 07:14:22 crc kubenswrapper[4809]: E1124 07:14:22.457286 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c0203aa-9e75-4bf4-8660-68f27b1b794b" containerName="nova-api-log" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.457300 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c0203aa-9e75-4bf4-8660-68f27b1b794b" containerName="nova-api-log" Nov 24 07:14:22 crc kubenswrapper[4809]: E1124 07:14:22.457317 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c0203aa-9e75-4bf4-8660-68f27b1b794b" containerName="nova-api-api" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.457324 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c0203aa-9e75-4bf4-8660-68f27b1b794b" containerName="nova-api-api" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.457492 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c0203aa-9e75-4bf4-8660-68f27b1b794b" containerName="nova-api-api" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.457514 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c0203aa-9e75-4bf4-8660-68f27b1b794b" containerName="nova-api-log" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.458465 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.464337 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.464477 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.464631 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.466694 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.620080 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02e88e29-10b0-42fb-96d3-f562aa48f361-logs\") pod \"nova-api-0\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " pod="openstack/nova-api-0" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.620434 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-internal-tls-certs\") pod \"nova-api-0\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " pod="openstack/nova-api-0" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.620556 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-config-data\") pod \"nova-api-0\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " pod="openstack/nova-api-0" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.620779 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9z5f\" (UniqueName: \"kubernetes.io/projected/02e88e29-10b0-42fb-96d3-f562aa48f361-kube-api-access-c9z5f\") pod \"nova-api-0\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " pod="openstack/nova-api-0" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.620948 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " pod="openstack/nova-api-0" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.621009 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-public-tls-certs\") pod \"nova-api-0\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " pod="openstack/nova-api-0" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.722678 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-public-tls-certs\") pod \"nova-api-0\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " pod="openstack/nova-api-0" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.722714 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " pod="openstack/nova-api-0" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.722774 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02e88e29-10b0-42fb-96d3-f562aa48f361-logs\") pod \"nova-api-0\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " pod="openstack/nova-api-0" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.722808 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-internal-tls-certs\") pod \"nova-api-0\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " pod="openstack/nova-api-0" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.722843 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-config-data\") pod \"nova-api-0\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " pod="openstack/nova-api-0" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.722923 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9z5f\" (UniqueName: \"kubernetes.io/projected/02e88e29-10b0-42fb-96d3-f562aa48f361-kube-api-access-c9z5f\") pod \"nova-api-0\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " pod="openstack/nova-api-0" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.723588 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02e88e29-10b0-42fb-96d3-f562aa48f361-logs\") pod \"nova-api-0\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " pod="openstack/nova-api-0" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.728763 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-public-tls-certs\") pod \"nova-api-0\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " pod="openstack/nova-api-0" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.728947 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-internal-tls-certs\") pod \"nova-api-0\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " pod="openstack/nova-api-0" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.729034 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-config-data\") pod \"nova-api-0\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " pod="openstack/nova-api-0" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.729061 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " pod="openstack/nova-api-0" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.758421 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9z5f\" (UniqueName: \"kubernetes.io/projected/02e88e29-10b0-42fb-96d3-f562aa48f361-kube-api-access-c9z5f\") pod \"nova-api-0\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " pod="openstack/nova-api-0" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.778035 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:14:22 crc kubenswrapper[4809]: I1124 07:14:22.904836 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c0203aa-9e75-4bf4-8660-68f27b1b794b" path="/var/lib/kubelet/pods/2c0203aa-9e75-4bf4-8660-68f27b1b794b/volumes" Nov 24 07:14:23 crc kubenswrapper[4809]: W1124 07:14:23.256368 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02e88e29_10b0_42fb_96d3_f562aa48f361.slice/crio-13a40a4f800906df01f104c13146716ecc1e7c8570ebadc2b2043b51d024e664 WatchSource:0}: Error finding container 13a40a4f800906df01f104c13146716ecc1e7c8570ebadc2b2043b51d024e664: Status 404 returned error can't find the container with id 13a40a4f800906df01f104c13146716ecc1e7c8570ebadc2b2043b51d024e664 Nov 24 07:14:23 crc kubenswrapper[4809]: I1124 07:14:23.262885 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:14:23 crc kubenswrapper[4809]: I1124 07:14:23.373568 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"02e88e29-10b0-42fb-96d3-f562aa48f361","Type":"ContainerStarted","Data":"13a40a4f800906df01f104c13146716ecc1e7c8570ebadc2b2043b51d024e664"} Nov 24 07:14:23 crc kubenswrapper[4809]: I1124 07:14:23.761065 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:14:24 crc kubenswrapper[4809]: I1124 07:14:24.386445 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52","Type":"ContainerStarted","Data":"4635862bdee21ba43a65de721027f25742f5c6cfb700fddab875b3f711217246"} Nov 24 07:14:24 crc kubenswrapper[4809]: I1124 07:14:24.388438 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"02e88e29-10b0-42fb-96d3-f562aa48f361","Type":"ContainerStarted","Data":"c44af3db3c24c41298e54d466fee3a1916bde42927dd20216ef9ac046f32783e"} Nov 24 07:14:24 crc kubenswrapper[4809]: I1124 07:14:24.388564 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"02e88e29-10b0-42fb-96d3-f562aa48f361","Type":"ContainerStarted","Data":"b95add818e2dacbe46ab6681b06ab893d4b986180b8af6f266e102d9e3a57267"} Nov 24 07:14:24 crc kubenswrapper[4809]: I1124 07:14:24.414695 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.414677771 podStartE2EDuration="2.414677771s" podCreationTimestamp="2025-11-24 07:14:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:14:24.402777836 +0000 UTC m=+1164.303369541" watchObservedRunningTime="2025-11-24 07:14:24.414677771 +0000 UTC m=+1164.315269476" Nov 24 07:14:25 crc kubenswrapper[4809]: I1124 07:14:25.747902 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:14:25 crc kubenswrapper[4809]: I1124 07:14:25.841264 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-jw4ks"] Nov 24 07:14:25 crc kubenswrapper[4809]: I1124 07:14:25.841528 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" podUID="9feafcbd-4cea-47dc-b276-6878ad28e569" containerName="dnsmasq-dns" containerID="cri-o://947d9b54a8eba14ac0782c84ff11b1e09808c99629b722515f0308ae29545aff" gracePeriod=10 Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.365785 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.419500 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52","Type":"ContainerStarted","Data":"a526abb06386dc7de473085fc33ff9bf96a8bce761e99043eb43e079919267a1"} Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.419825 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.419613 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" containerName="ceilometer-central-agent" containerID="cri-o://6f443b63c63f414412d4726e5018d4498ace22da416c81c9def3225bf2cbcbc3" gracePeriod=30 Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.419923 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" containerName="proxy-httpd" containerID="cri-o://a526abb06386dc7de473085fc33ff9bf96a8bce761e99043eb43e079919267a1" gracePeriod=30 Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.419945 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" containerName="ceilometer-notification-agent" containerID="cri-o://0e5e353e5533e9b43aa58a962905998869290a3459004a4dc8e03bf9f8ac619c" gracePeriod=30 Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.420019 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" containerName="sg-core" containerID="cri-o://4635862bdee21ba43a65de721027f25742f5c6cfb700fddab875b3f711217246" gracePeriod=30 Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.422026 4809 generic.go:334] "Generic (PLEG): container finished" podID="9feafcbd-4cea-47dc-b276-6878ad28e569" containerID="947d9b54a8eba14ac0782c84ff11b1e09808c99629b722515f0308ae29545aff" exitCode=0 Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.422149 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" event={"ID":"9feafcbd-4cea-47dc-b276-6878ad28e569","Type":"ContainerDied","Data":"947d9b54a8eba14ac0782c84ff11b1e09808c99629b722515f0308ae29545aff"} Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.422241 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" event={"ID":"9feafcbd-4cea-47dc-b276-6878ad28e569","Type":"ContainerDied","Data":"ecdb8eadfaf9a4e540d1e16b5a977e33073674718337a3d21e2ce610bf76e862"} Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.422312 4809 scope.go:117] "RemoveContainer" containerID="947d9b54a8eba14ac0782c84ff11b1e09808c99629b722515f0308ae29545aff" Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.422487 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-jw4ks" Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.440858 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.447190767 podStartE2EDuration="7.440840398s" podCreationTimestamp="2025-11-24 07:14:19 +0000 UTC" firstStartedPulling="2025-11-24 07:14:20.217675811 +0000 UTC m=+1160.118267516" lastFinishedPulling="2025-11-24 07:14:25.211325442 +0000 UTC m=+1165.111917147" observedRunningTime="2025-11-24 07:14:26.437441539 +0000 UTC m=+1166.338033264" watchObservedRunningTime="2025-11-24 07:14:26.440840398 +0000 UTC m=+1166.341432113" Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.449150 4809 scope.go:117] "RemoveContainer" containerID="cffd6f156e3e5b8f3adb7309fb68d474c847d84da55d7163825a8bd94aacc6b7" Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.478737 4809 scope.go:117] "RemoveContainer" containerID="947d9b54a8eba14ac0782c84ff11b1e09808c99629b722515f0308ae29545aff" Nov 24 07:14:26 crc kubenswrapper[4809]: E1124 07:14:26.479330 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"947d9b54a8eba14ac0782c84ff11b1e09808c99629b722515f0308ae29545aff\": container with ID starting with 947d9b54a8eba14ac0782c84ff11b1e09808c99629b722515f0308ae29545aff not found: ID does not exist" containerID="947d9b54a8eba14ac0782c84ff11b1e09808c99629b722515f0308ae29545aff" Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.479369 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"947d9b54a8eba14ac0782c84ff11b1e09808c99629b722515f0308ae29545aff"} err="failed to get container status \"947d9b54a8eba14ac0782c84ff11b1e09808c99629b722515f0308ae29545aff\": rpc error: code = NotFound desc = could not find container \"947d9b54a8eba14ac0782c84ff11b1e09808c99629b722515f0308ae29545aff\": container with ID starting with 947d9b54a8eba14ac0782c84ff11b1e09808c99629b722515f0308ae29545aff not found: ID does not exist" Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.479427 4809 scope.go:117] "RemoveContainer" containerID="cffd6f156e3e5b8f3adb7309fb68d474c847d84da55d7163825a8bd94aacc6b7" Nov 24 07:14:26 crc kubenswrapper[4809]: E1124 07:14:26.480116 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cffd6f156e3e5b8f3adb7309fb68d474c847d84da55d7163825a8bd94aacc6b7\": container with ID starting with cffd6f156e3e5b8f3adb7309fb68d474c847d84da55d7163825a8bd94aacc6b7 not found: ID does not exist" containerID="cffd6f156e3e5b8f3adb7309fb68d474c847d84da55d7163825a8bd94aacc6b7" Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.480156 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cffd6f156e3e5b8f3adb7309fb68d474c847d84da55d7163825a8bd94aacc6b7"} err="failed to get container status \"cffd6f156e3e5b8f3adb7309fb68d474c847d84da55d7163825a8bd94aacc6b7\": rpc error: code = NotFound desc = could not find container \"cffd6f156e3e5b8f3adb7309fb68d474c847d84da55d7163825a8bd94aacc6b7\": container with ID starting with cffd6f156e3e5b8f3adb7309fb68d474c847d84da55d7163825a8bd94aacc6b7 not found: ID does not exist" Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.514578 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-ovsdbserver-sb\") pod \"9feafcbd-4cea-47dc-b276-6878ad28e569\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.514635 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-dns-swift-storage-0\") pod \"9feafcbd-4cea-47dc-b276-6878ad28e569\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.514787 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-dns-svc\") pod \"9feafcbd-4cea-47dc-b276-6878ad28e569\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.514843 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbjsp\" (UniqueName: \"kubernetes.io/projected/9feafcbd-4cea-47dc-b276-6878ad28e569-kube-api-access-fbjsp\") pod \"9feafcbd-4cea-47dc-b276-6878ad28e569\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.514862 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-config\") pod \"9feafcbd-4cea-47dc-b276-6878ad28e569\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.514956 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-ovsdbserver-nb\") pod \"9feafcbd-4cea-47dc-b276-6878ad28e569\" (UID: \"9feafcbd-4cea-47dc-b276-6878ad28e569\") " Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.521290 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9feafcbd-4cea-47dc-b276-6878ad28e569-kube-api-access-fbjsp" (OuterVolumeSpecName: "kube-api-access-fbjsp") pod "9feafcbd-4cea-47dc-b276-6878ad28e569" (UID: "9feafcbd-4cea-47dc-b276-6878ad28e569"). InnerVolumeSpecName "kube-api-access-fbjsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.571587 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-config" (OuterVolumeSpecName: "config") pod "9feafcbd-4cea-47dc-b276-6878ad28e569" (UID: "9feafcbd-4cea-47dc-b276-6878ad28e569"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.571705 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9feafcbd-4cea-47dc-b276-6878ad28e569" (UID: "9feafcbd-4cea-47dc-b276-6878ad28e569"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.574480 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9feafcbd-4cea-47dc-b276-6878ad28e569" (UID: "9feafcbd-4cea-47dc-b276-6878ad28e569"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.577405 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9feafcbd-4cea-47dc-b276-6878ad28e569" (UID: "9feafcbd-4cea-47dc-b276-6878ad28e569"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.584374 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9feafcbd-4cea-47dc-b276-6878ad28e569" (UID: "9feafcbd-4cea-47dc-b276-6878ad28e569"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.617094 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.617134 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbjsp\" (UniqueName: \"kubernetes.io/projected/9feafcbd-4cea-47dc-b276-6878ad28e569-kube-api-access-fbjsp\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.617150 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.617162 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.617174 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.617186 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9feafcbd-4cea-47dc-b276-6878ad28e569-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.756352 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-jw4ks"] Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.763218 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-jw4ks"] Nov 24 07:14:26 crc kubenswrapper[4809]: I1124 07:14:26.903106 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9feafcbd-4cea-47dc-b276-6878ad28e569" path="/var/lib/kubelet/pods/9feafcbd-4cea-47dc-b276-6878ad28e569/volumes" Nov 24 07:14:27 crc kubenswrapper[4809]: I1124 07:14:27.433173 4809 generic.go:334] "Generic (PLEG): container finished" podID="d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" containerID="a526abb06386dc7de473085fc33ff9bf96a8bce761e99043eb43e079919267a1" exitCode=0 Nov 24 07:14:27 crc kubenswrapper[4809]: I1124 07:14:27.433470 4809 generic.go:334] "Generic (PLEG): container finished" podID="d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" containerID="4635862bdee21ba43a65de721027f25742f5c6cfb700fddab875b3f711217246" exitCode=2 Nov 24 07:14:27 crc kubenswrapper[4809]: I1124 07:14:27.433480 4809 generic.go:334] "Generic (PLEG): container finished" podID="d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" containerID="0e5e353e5533e9b43aa58a962905998869290a3459004a4dc8e03bf9f8ac619c" exitCode=0 Nov 24 07:14:27 crc kubenswrapper[4809]: I1124 07:14:27.433233 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52","Type":"ContainerDied","Data":"a526abb06386dc7de473085fc33ff9bf96a8bce761e99043eb43e079919267a1"} Nov 24 07:14:27 crc kubenswrapper[4809]: I1124 07:14:27.433546 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52","Type":"ContainerDied","Data":"4635862bdee21ba43a65de721027f25742f5c6cfb700fddab875b3f711217246"} Nov 24 07:14:27 crc kubenswrapper[4809]: I1124 07:14:27.433564 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52","Type":"ContainerDied","Data":"0e5e353e5533e9b43aa58a962905998869290a3459004a4dc8e03bf9f8ac619c"} Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.054279 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.246646 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-config-data\") pod \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.246733 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-run-httpd\") pod \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.246754 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-ceilometer-tls-certs\") pod \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.246793 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-log-httpd\") pod \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.246837 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-sg-core-conf-yaml\") pod \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.246860 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpx4w\" (UniqueName: \"kubernetes.io/projected/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-kube-api-access-bpx4w\") pod \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.246892 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-combined-ca-bundle\") pod \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.246961 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-scripts\") pod \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\" (UID: \"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52\") " Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.247601 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" (UID: "d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.247617 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" (UID: "d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.248156 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.248177 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.253320 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-scripts" (OuterVolumeSpecName: "scripts") pod "d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" (UID: "d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.253553 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-kube-api-access-bpx4w" (OuterVolumeSpecName: "kube-api-access-bpx4w") pod "d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" (UID: "d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52"). InnerVolumeSpecName "kube-api-access-bpx4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.281331 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" (UID: "d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.307354 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" (UID: "d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.333936 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" (UID: "d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.345810 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-config-data" (OuterVolumeSpecName: "config-data") pod "d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" (UID: "d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.349988 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.350014 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.350024 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.350035 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpx4w\" (UniqueName: \"kubernetes.io/projected/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-kube-api-access-bpx4w\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.350044 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.350051 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.447801 4809 generic.go:334] "Generic (PLEG): container finished" podID="d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" containerID="6f443b63c63f414412d4726e5018d4498ace22da416c81c9def3225bf2cbcbc3" exitCode=0 Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.447841 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52","Type":"ContainerDied","Data":"6f443b63c63f414412d4726e5018d4498ace22da416c81c9def3225bf2cbcbc3"} Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.447866 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52","Type":"ContainerDied","Data":"f23ab21a2611f62431ae3cceae0365a16a05e7bb981b9a13ba8c5a3520088ee9"} Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.447886 4809 scope.go:117] "RemoveContainer" containerID="a526abb06386dc7de473085fc33ff9bf96a8bce761e99043eb43e079919267a1" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.448022 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.480505 4809 scope.go:117] "RemoveContainer" containerID="4635862bdee21ba43a65de721027f25742f5c6cfb700fddab875b3f711217246" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.500485 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.513405 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.522131 4809 scope.go:117] "RemoveContainer" containerID="0e5e353e5533e9b43aa58a962905998869290a3459004a4dc8e03bf9f8ac619c" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.522288 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:14:28 crc kubenswrapper[4809]: E1124 07:14:28.522715 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" containerName="proxy-httpd" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.522734 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" containerName="proxy-httpd" Nov 24 07:14:28 crc kubenswrapper[4809]: E1124 07:14:28.522746 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9feafcbd-4cea-47dc-b276-6878ad28e569" containerName="init" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.522755 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9feafcbd-4cea-47dc-b276-6878ad28e569" containerName="init" Nov 24 07:14:28 crc kubenswrapper[4809]: E1124 07:14:28.522778 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9feafcbd-4cea-47dc-b276-6878ad28e569" containerName="dnsmasq-dns" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.522786 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9feafcbd-4cea-47dc-b276-6878ad28e569" containerName="dnsmasq-dns" Nov 24 07:14:28 crc kubenswrapper[4809]: E1124 07:14:28.522811 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" containerName="ceilometer-notification-agent" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.522818 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" containerName="ceilometer-notification-agent" Nov 24 07:14:28 crc kubenswrapper[4809]: E1124 07:14:28.522830 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" containerName="ceilometer-central-agent" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.522839 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" containerName="ceilometer-central-agent" Nov 24 07:14:28 crc kubenswrapper[4809]: E1124 07:14:28.522853 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" containerName="sg-core" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.522859 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" containerName="sg-core" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.523097 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" containerName="ceilometer-central-agent" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.523116 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" containerName="sg-core" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.523135 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" containerName="proxy-httpd" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.523144 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" containerName="ceilometer-notification-agent" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.523151 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9feafcbd-4cea-47dc-b276-6878ad28e569" containerName="dnsmasq-dns" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.525164 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.528047 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.529194 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.529358 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.540180 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.553857 4809 scope.go:117] "RemoveContainer" containerID="6f443b63c63f414412d4726e5018d4498ace22da416c81c9def3225bf2cbcbc3" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.572697 4809 scope.go:117] "RemoveContainer" containerID="a526abb06386dc7de473085fc33ff9bf96a8bce761e99043eb43e079919267a1" Nov 24 07:14:28 crc kubenswrapper[4809]: E1124 07:14:28.573194 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a526abb06386dc7de473085fc33ff9bf96a8bce761e99043eb43e079919267a1\": container with ID starting with a526abb06386dc7de473085fc33ff9bf96a8bce761e99043eb43e079919267a1 not found: ID does not exist" containerID="a526abb06386dc7de473085fc33ff9bf96a8bce761e99043eb43e079919267a1" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.573353 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a526abb06386dc7de473085fc33ff9bf96a8bce761e99043eb43e079919267a1"} err="failed to get container status \"a526abb06386dc7de473085fc33ff9bf96a8bce761e99043eb43e079919267a1\": rpc error: code = NotFound desc = could not find container \"a526abb06386dc7de473085fc33ff9bf96a8bce761e99043eb43e079919267a1\": container with ID starting with a526abb06386dc7de473085fc33ff9bf96a8bce761e99043eb43e079919267a1 not found: ID does not exist" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.573487 4809 scope.go:117] "RemoveContainer" containerID="4635862bdee21ba43a65de721027f25742f5c6cfb700fddab875b3f711217246" Nov 24 07:14:28 crc kubenswrapper[4809]: E1124 07:14:28.573930 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4635862bdee21ba43a65de721027f25742f5c6cfb700fddab875b3f711217246\": container with ID starting with 4635862bdee21ba43a65de721027f25742f5c6cfb700fddab875b3f711217246 not found: ID does not exist" containerID="4635862bdee21ba43a65de721027f25742f5c6cfb700fddab875b3f711217246" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.573990 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4635862bdee21ba43a65de721027f25742f5c6cfb700fddab875b3f711217246"} err="failed to get container status \"4635862bdee21ba43a65de721027f25742f5c6cfb700fddab875b3f711217246\": rpc error: code = NotFound desc = could not find container \"4635862bdee21ba43a65de721027f25742f5c6cfb700fddab875b3f711217246\": container with ID starting with 4635862bdee21ba43a65de721027f25742f5c6cfb700fddab875b3f711217246 not found: ID does not exist" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.574019 4809 scope.go:117] "RemoveContainer" containerID="0e5e353e5533e9b43aa58a962905998869290a3459004a4dc8e03bf9f8ac619c" Nov 24 07:14:28 crc kubenswrapper[4809]: E1124 07:14:28.574393 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e5e353e5533e9b43aa58a962905998869290a3459004a4dc8e03bf9f8ac619c\": container with ID starting with 0e5e353e5533e9b43aa58a962905998869290a3459004a4dc8e03bf9f8ac619c not found: ID does not exist" containerID="0e5e353e5533e9b43aa58a962905998869290a3459004a4dc8e03bf9f8ac619c" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.574531 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e5e353e5533e9b43aa58a962905998869290a3459004a4dc8e03bf9f8ac619c"} err="failed to get container status \"0e5e353e5533e9b43aa58a962905998869290a3459004a4dc8e03bf9f8ac619c\": rpc error: code = NotFound desc = could not find container \"0e5e353e5533e9b43aa58a962905998869290a3459004a4dc8e03bf9f8ac619c\": container with ID starting with 0e5e353e5533e9b43aa58a962905998869290a3459004a4dc8e03bf9f8ac619c not found: ID does not exist" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.574650 4809 scope.go:117] "RemoveContainer" containerID="6f443b63c63f414412d4726e5018d4498ace22da416c81c9def3225bf2cbcbc3" Nov 24 07:14:28 crc kubenswrapper[4809]: E1124 07:14:28.575039 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f443b63c63f414412d4726e5018d4498ace22da416c81c9def3225bf2cbcbc3\": container with ID starting with 6f443b63c63f414412d4726e5018d4498ace22da416c81c9def3225bf2cbcbc3 not found: ID does not exist" containerID="6f443b63c63f414412d4726e5018d4498ace22da416c81c9def3225bf2cbcbc3" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.575079 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f443b63c63f414412d4726e5018d4498ace22da416c81c9def3225bf2cbcbc3"} err="failed to get container status \"6f443b63c63f414412d4726e5018d4498ace22da416c81c9def3225bf2cbcbc3\": rpc error: code = NotFound desc = could not find container \"6f443b63c63f414412d4726e5018d4498ace22da416c81c9def3225bf2cbcbc3\": container with ID starting with 6f443b63c63f414412d4726e5018d4498ace22da416c81c9def3225bf2cbcbc3 not found: ID does not exist" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.653691 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96597dff-85f6-4758-a2e5-27c5b01bf34a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.653741 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96597dff-85f6-4758-a2e5-27c5b01bf34a-config-data\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.653761 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96597dff-85f6-4758-a2e5-27c5b01bf34a-log-httpd\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.654561 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/96597dff-85f6-4758-a2e5-27c5b01bf34a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.654737 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96597dff-85f6-4758-a2e5-27c5b01bf34a-scripts\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.654859 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96597dff-85f6-4758-a2e5-27c5b01bf34a-run-httpd\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.654983 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/96597dff-85f6-4758-a2e5-27c5b01bf34a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.655169 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddp2b\" (UniqueName: \"kubernetes.io/projected/96597dff-85f6-4758-a2e5-27c5b01bf34a-kube-api-access-ddp2b\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.756435 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/96597dff-85f6-4758-a2e5-27c5b01bf34a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.756551 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddp2b\" (UniqueName: \"kubernetes.io/projected/96597dff-85f6-4758-a2e5-27c5b01bf34a-kube-api-access-ddp2b\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.756654 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96597dff-85f6-4758-a2e5-27c5b01bf34a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.756686 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96597dff-85f6-4758-a2e5-27c5b01bf34a-config-data\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.756708 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96597dff-85f6-4758-a2e5-27c5b01bf34a-log-httpd\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.756772 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/96597dff-85f6-4758-a2e5-27c5b01bf34a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.756819 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96597dff-85f6-4758-a2e5-27c5b01bf34a-scripts\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.756857 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96597dff-85f6-4758-a2e5-27c5b01bf34a-run-httpd\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.757231 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96597dff-85f6-4758-a2e5-27c5b01bf34a-log-httpd\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.757371 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96597dff-85f6-4758-a2e5-27c5b01bf34a-run-httpd\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.760841 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.761123 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96597dff-85f6-4758-a2e5-27c5b01bf34a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.761300 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/96597dff-85f6-4758-a2e5-27c5b01bf34a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.761673 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/96597dff-85f6-4758-a2e5-27c5b01bf34a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.761860 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96597dff-85f6-4758-a2e5-27c5b01bf34a-scripts\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.762201 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96597dff-85f6-4758-a2e5-27c5b01bf34a-config-data\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.774624 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddp2b\" (UniqueName: \"kubernetes.io/projected/96597dff-85f6-4758-a2e5-27c5b01bf34a-kube-api-access-ddp2b\") pod \"ceilometer-0\" (UID: \"96597dff-85f6-4758-a2e5-27c5b01bf34a\") " pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.793269 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.858228 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:14:28 crc kubenswrapper[4809]: I1124 07:14:28.903874 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52" path="/var/lib/kubelet/pods/d0f8ff3f-5d13-43db-8f4f-d4d1cc9bee52/volumes" Nov 24 07:14:29 crc kubenswrapper[4809]: I1124 07:14:29.315674 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:14:29 crc kubenswrapper[4809]: W1124 07:14:29.320592 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96597dff_85f6_4758_a2e5_27c5b01bf34a.slice/crio-07baa30e2d36451f73b0d41237c1da58d6bd83a2bc8b93c2fb988849c5620cd5 WatchSource:0}: Error finding container 07baa30e2d36451f73b0d41237c1da58d6bd83a2bc8b93c2fb988849c5620cd5: Status 404 returned error can't find the container with id 07baa30e2d36451f73b0d41237c1da58d6bd83a2bc8b93c2fb988849c5620cd5 Nov 24 07:14:29 crc kubenswrapper[4809]: I1124 07:14:29.460313 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96597dff-85f6-4758-a2e5-27c5b01bf34a","Type":"ContainerStarted","Data":"07baa30e2d36451f73b0d41237c1da58d6bd83a2bc8b93c2fb988849c5620cd5"} Nov 24 07:14:29 crc kubenswrapper[4809]: I1124 07:14:29.490192 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:14:29 crc kubenswrapper[4809]: I1124 07:14:29.632939 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-9v7n9"] Nov 24 07:14:29 crc kubenswrapper[4809]: I1124 07:14:29.634197 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9v7n9" Nov 24 07:14:29 crc kubenswrapper[4809]: I1124 07:14:29.636316 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 24 07:14:29 crc kubenswrapper[4809]: I1124 07:14:29.646617 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-9v7n9"] Nov 24 07:14:29 crc kubenswrapper[4809]: I1124 07:14:29.648785 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 24 07:14:29 crc kubenswrapper[4809]: I1124 07:14:29.673860 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97266dae-027e-469f-8934-298e38fbde54-config-data\") pod \"nova-cell1-cell-mapping-9v7n9\" (UID: \"97266dae-027e-469f-8934-298e38fbde54\") " pod="openstack/nova-cell1-cell-mapping-9v7n9" Nov 24 07:14:29 crc kubenswrapper[4809]: I1124 07:14:29.674160 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfmvw\" (UniqueName: \"kubernetes.io/projected/97266dae-027e-469f-8934-298e38fbde54-kube-api-access-bfmvw\") pod \"nova-cell1-cell-mapping-9v7n9\" (UID: \"97266dae-027e-469f-8934-298e38fbde54\") " pod="openstack/nova-cell1-cell-mapping-9v7n9" Nov 24 07:14:29 crc kubenswrapper[4809]: I1124 07:14:29.674208 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97266dae-027e-469f-8934-298e38fbde54-scripts\") pod \"nova-cell1-cell-mapping-9v7n9\" (UID: \"97266dae-027e-469f-8934-298e38fbde54\") " pod="openstack/nova-cell1-cell-mapping-9v7n9" Nov 24 07:14:29 crc kubenswrapper[4809]: I1124 07:14:29.674228 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97266dae-027e-469f-8934-298e38fbde54-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9v7n9\" (UID: \"97266dae-027e-469f-8934-298e38fbde54\") " pod="openstack/nova-cell1-cell-mapping-9v7n9" Nov 24 07:14:29 crc kubenswrapper[4809]: I1124 07:14:29.775680 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfmvw\" (UniqueName: \"kubernetes.io/projected/97266dae-027e-469f-8934-298e38fbde54-kube-api-access-bfmvw\") pod \"nova-cell1-cell-mapping-9v7n9\" (UID: \"97266dae-027e-469f-8934-298e38fbde54\") " pod="openstack/nova-cell1-cell-mapping-9v7n9" Nov 24 07:14:29 crc kubenswrapper[4809]: I1124 07:14:29.775729 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97266dae-027e-469f-8934-298e38fbde54-scripts\") pod \"nova-cell1-cell-mapping-9v7n9\" (UID: \"97266dae-027e-469f-8934-298e38fbde54\") " pod="openstack/nova-cell1-cell-mapping-9v7n9" Nov 24 07:14:29 crc kubenswrapper[4809]: I1124 07:14:29.775747 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97266dae-027e-469f-8934-298e38fbde54-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9v7n9\" (UID: \"97266dae-027e-469f-8934-298e38fbde54\") " pod="openstack/nova-cell1-cell-mapping-9v7n9" Nov 24 07:14:29 crc kubenswrapper[4809]: I1124 07:14:29.775781 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97266dae-027e-469f-8934-298e38fbde54-config-data\") pod \"nova-cell1-cell-mapping-9v7n9\" (UID: \"97266dae-027e-469f-8934-298e38fbde54\") " pod="openstack/nova-cell1-cell-mapping-9v7n9" Nov 24 07:14:29 crc kubenswrapper[4809]: I1124 07:14:29.781160 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97266dae-027e-469f-8934-298e38fbde54-scripts\") pod \"nova-cell1-cell-mapping-9v7n9\" (UID: \"97266dae-027e-469f-8934-298e38fbde54\") " pod="openstack/nova-cell1-cell-mapping-9v7n9" Nov 24 07:14:29 crc kubenswrapper[4809]: I1124 07:14:29.781464 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97266dae-027e-469f-8934-298e38fbde54-config-data\") pod \"nova-cell1-cell-mapping-9v7n9\" (UID: \"97266dae-027e-469f-8934-298e38fbde54\") " pod="openstack/nova-cell1-cell-mapping-9v7n9" Nov 24 07:14:29 crc kubenswrapper[4809]: I1124 07:14:29.782628 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97266dae-027e-469f-8934-298e38fbde54-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9v7n9\" (UID: \"97266dae-027e-469f-8934-298e38fbde54\") " pod="openstack/nova-cell1-cell-mapping-9v7n9" Nov 24 07:14:29 crc kubenswrapper[4809]: I1124 07:14:29.791497 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfmvw\" (UniqueName: \"kubernetes.io/projected/97266dae-027e-469f-8934-298e38fbde54-kube-api-access-bfmvw\") pod \"nova-cell1-cell-mapping-9v7n9\" (UID: \"97266dae-027e-469f-8934-298e38fbde54\") " pod="openstack/nova-cell1-cell-mapping-9v7n9" Nov 24 07:14:29 crc kubenswrapper[4809]: I1124 07:14:29.962454 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9v7n9" Nov 24 07:14:30 crc kubenswrapper[4809]: I1124 07:14:30.378528 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-9v7n9"] Nov 24 07:14:30 crc kubenswrapper[4809]: W1124 07:14:30.378895 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97266dae_027e_469f_8934_298e38fbde54.slice/crio-df2249c1cb67a9f667839ce2264a2254110fdf136a5838d7c88005a5e4021b0a WatchSource:0}: Error finding container df2249c1cb67a9f667839ce2264a2254110fdf136a5838d7c88005a5e4021b0a: Status 404 returned error can't find the container with id df2249c1cb67a9f667839ce2264a2254110fdf136a5838d7c88005a5e4021b0a Nov 24 07:14:30 crc kubenswrapper[4809]: I1124 07:14:30.468139 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9v7n9" event={"ID":"97266dae-027e-469f-8934-298e38fbde54","Type":"ContainerStarted","Data":"df2249c1cb67a9f667839ce2264a2254110fdf136a5838d7c88005a5e4021b0a"} Nov 24 07:14:30 crc kubenswrapper[4809]: I1124 07:14:30.469640 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96597dff-85f6-4758-a2e5-27c5b01bf34a","Type":"ContainerStarted","Data":"4b21b5577db1a63ceae482ca843bbd329fa77703adcd8c29f85c8ac74eb2b3f2"} Nov 24 07:14:31 crc kubenswrapper[4809]: I1124 07:14:31.479879 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96597dff-85f6-4758-a2e5-27c5b01bf34a","Type":"ContainerStarted","Data":"e43da8531a70c8bf7976bec23be5d9f207fda65fd28a0220471f13eccf895d3c"} Nov 24 07:14:31 crc kubenswrapper[4809]: I1124 07:14:31.482213 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9v7n9" event={"ID":"97266dae-027e-469f-8934-298e38fbde54","Type":"ContainerStarted","Data":"212b977fdd911e9454a6815429587513ec64237aaa0ddea848608ae8ce5ca28d"} Nov 24 07:14:32 crc kubenswrapper[4809]: I1124 07:14:32.491681 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96597dff-85f6-4758-a2e5-27c5b01bf34a","Type":"ContainerStarted","Data":"a401c0f2a9b7694b633170194e4f47a6d337eaddb029d6a9458aafe1782ed2e8"} Nov 24 07:14:32 crc kubenswrapper[4809]: I1124 07:14:32.782104 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 07:14:32 crc kubenswrapper[4809]: I1124 07:14:32.782166 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 07:14:33 crc kubenswrapper[4809]: I1124 07:14:33.862324 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="02e88e29-10b0-42fb-96d3-f562aa48f361" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 07:14:33 crc kubenswrapper[4809]: I1124 07:14:33.862347 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="02e88e29-10b0-42fb-96d3-f562aa48f361" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 07:14:34 crc kubenswrapper[4809]: I1124 07:14:34.516127 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96597dff-85f6-4758-a2e5-27c5b01bf34a","Type":"ContainerStarted","Data":"f076f11c5de59cd4d061ee23a3d9745e89ee3ec398c1ba7dc0e79a4048ade574"} Nov 24 07:14:34 crc kubenswrapper[4809]: I1124 07:14:34.517709 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 07:14:34 crc kubenswrapper[4809]: I1124 07:14:34.543645 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.103958545 podStartE2EDuration="6.543630569s" podCreationTimestamp="2025-11-24 07:14:28 +0000 UTC" firstStartedPulling="2025-11-24 07:14:29.324706776 +0000 UTC m=+1169.225298481" lastFinishedPulling="2025-11-24 07:14:33.7643788 +0000 UTC m=+1173.664970505" observedRunningTime="2025-11-24 07:14:34.541343378 +0000 UTC m=+1174.441935083" watchObservedRunningTime="2025-11-24 07:14:34.543630569 +0000 UTC m=+1174.444222274" Nov 24 07:14:34 crc kubenswrapper[4809]: I1124 07:14:34.546424 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-9v7n9" podStartSLOduration=5.546413943 podStartE2EDuration="5.546413943s" podCreationTimestamp="2025-11-24 07:14:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:14:31.50728553 +0000 UTC m=+1171.407877225" watchObservedRunningTime="2025-11-24 07:14:34.546413943 +0000 UTC m=+1174.447005648" Nov 24 07:14:35 crc kubenswrapper[4809]: I1124 07:14:35.531460 4809 generic.go:334] "Generic (PLEG): container finished" podID="97266dae-027e-469f-8934-298e38fbde54" containerID="212b977fdd911e9454a6815429587513ec64237aaa0ddea848608ae8ce5ca28d" exitCode=0 Nov 24 07:14:35 crc kubenswrapper[4809]: I1124 07:14:35.531583 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9v7n9" event={"ID":"97266dae-027e-469f-8934-298e38fbde54","Type":"ContainerDied","Data":"212b977fdd911e9454a6815429587513ec64237aaa0ddea848608ae8ce5ca28d"} Nov 24 07:14:36 crc kubenswrapper[4809]: I1124 07:14:36.870958 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9v7n9" Nov 24 07:14:37 crc kubenswrapper[4809]: I1124 07:14:37.025145 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97266dae-027e-469f-8934-298e38fbde54-config-data\") pod \"97266dae-027e-469f-8934-298e38fbde54\" (UID: \"97266dae-027e-469f-8934-298e38fbde54\") " Nov 24 07:14:37 crc kubenswrapper[4809]: I1124 07:14:37.025279 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97266dae-027e-469f-8934-298e38fbde54-scripts\") pod \"97266dae-027e-469f-8934-298e38fbde54\" (UID: \"97266dae-027e-469f-8934-298e38fbde54\") " Nov 24 07:14:37 crc kubenswrapper[4809]: I1124 07:14:37.025352 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97266dae-027e-469f-8934-298e38fbde54-combined-ca-bundle\") pod \"97266dae-027e-469f-8934-298e38fbde54\" (UID: \"97266dae-027e-469f-8934-298e38fbde54\") " Nov 24 07:14:37 crc kubenswrapper[4809]: I1124 07:14:37.025391 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfmvw\" (UniqueName: \"kubernetes.io/projected/97266dae-027e-469f-8934-298e38fbde54-kube-api-access-bfmvw\") pod \"97266dae-027e-469f-8934-298e38fbde54\" (UID: \"97266dae-027e-469f-8934-298e38fbde54\") " Nov 24 07:14:37 crc kubenswrapper[4809]: I1124 07:14:37.035107 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97266dae-027e-469f-8934-298e38fbde54-kube-api-access-bfmvw" (OuterVolumeSpecName: "kube-api-access-bfmvw") pod "97266dae-027e-469f-8934-298e38fbde54" (UID: "97266dae-027e-469f-8934-298e38fbde54"). InnerVolumeSpecName "kube-api-access-bfmvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:14:37 crc kubenswrapper[4809]: I1124 07:14:37.035100 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97266dae-027e-469f-8934-298e38fbde54-scripts" (OuterVolumeSpecName: "scripts") pod "97266dae-027e-469f-8934-298e38fbde54" (UID: "97266dae-027e-469f-8934-298e38fbde54"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:37 crc kubenswrapper[4809]: I1124 07:14:37.084846 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97266dae-027e-469f-8934-298e38fbde54-config-data" (OuterVolumeSpecName: "config-data") pod "97266dae-027e-469f-8934-298e38fbde54" (UID: "97266dae-027e-469f-8934-298e38fbde54"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:37 crc kubenswrapper[4809]: I1124 07:14:37.087685 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97266dae-027e-469f-8934-298e38fbde54-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "97266dae-027e-469f-8934-298e38fbde54" (UID: "97266dae-027e-469f-8934-298e38fbde54"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:37 crc kubenswrapper[4809]: I1124 07:14:37.128163 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97266dae-027e-469f-8934-298e38fbde54-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:37 crc kubenswrapper[4809]: I1124 07:14:37.128202 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfmvw\" (UniqueName: \"kubernetes.io/projected/97266dae-027e-469f-8934-298e38fbde54-kube-api-access-bfmvw\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:37 crc kubenswrapper[4809]: I1124 07:14:37.128218 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97266dae-027e-469f-8934-298e38fbde54-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:37 crc kubenswrapper[4809]: I1124 07:14:37.128229 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97266dae-027e-469f-8934-298e38fbde54-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:37 crc kubenswrapper[4809]: I1124 07:14:37.552932 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9v7n9" event={"ID":"97266dae-027e-469f-8934-298e38fbde54","Type":"ContainerDied","Data":"df2249c1cb67a9f667839ce2264a2254110fdf136a5838d7c88005a5e4021b0a"} Nov 24 07:14:37 crc kubenswrapper[4809]: I1124 07:14:37.553058 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df2249c1cb67a9f667839ce2264a2254110fdf136a5838d7c88005a5e4021b0a" Nov 24 07:14:37 crc kubenswrapper[4809]: I1124 07:14:37.553002 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9v7n9" Nov 24 07:14:37 crc kubenswrapper[4809]: I1124 07:14:37.652473 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:14:37 crc kubenswrapper[4809]: I1124 07:14:37.652920 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="02e88e29-10b0-42fb-96d3-f562aa48f361" containerName="nova-api-log" containerID="cri-o://b95add818e2dacbe46ab6681b06ab893d4b986180b8af6f266e102d9e3a57267" gracePeriod=30 Nov 24 07:14:37 crc kubenswrapper[4809]: I1124 07:14:37.653078 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="02e88e29-10b0-42fb-96d3-f562aa48f361" containerName="nova-api-api" containerID="cri-o://c44af3db3c24c41298e54d466fee3a1916bde42927dd20216ef9ac046f32783e" gracePeriod=30 Nov 24 07:14:37 crc kubenswrapper[4809]: I1124 07:14:37.712111 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:14:37 crc kubenswrapper[4809]: I1124 07:14:37.712399 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="111098ed-9d16-45e0-bb80-35f9383e327e" containerName="nova-scheduler-scheduler" containerID="cri-o://ffb9df9f2296363c66395fba7c7aa3efbc5800406624655d76c1096454276be2" gracePeriod=30 Nov 24 07:14:37 crc kubenswrapper[4809]: I1124 07:14:37.725300 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:14:37 crc kubenswrapper[4809]: I1124 07:14:37.725546 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1286d658-6187-4509-9e24-1e0d6fbcab1e" containerName="nova-metadata-log" containerID="cri-o://0a3e1f620cd8d80a95f532ec6d9f8cb9f3a36e6385c2d51b423c6d336e475487" gracePeriod=30 Nov 24 07:14:37 crc kubenswrapper[4809]: I1124 07:14:37.726039 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1286d658-6187-4509-9e24-1e0d6fbcab1e" containerName="nova-metadata-metadata" containerID="cri-o://8252b5e689178dcecb351ad3aeeb9896ac6dbaa4ca0ff97c73985ad51a08b07b" gracePeriod=30 Nov 24 07:14:38 crc kubenswrapper[4809]: E1124 07:14:38.420707 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ffb9df9f2296363c66395fba7c7aa3efbc5800406624655d76c1096454276be2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 07:14:38 crc kubenswrapper[4809]: E1124 07:14:38.425040 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ffb9df9f2296363c66395fba7c7aa3efbc5800406624655d76c1096454276be2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 07:14:38 crc kubenswrapper[4809]: E1124 07:14:38.426360 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ffb9df9f2296363c66395fba7c7aa3efbc5800406624655d76c1096454276be2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 07:14:38 crc kubenswrapper[4809]: E1124 07:14:38.426672 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="111098ed-9d16-45e0-bb80-35f9383e327e" containerName="nova-scheduler-scheduler" Nov 24 07:14:38 crc kubenswrapper[4809]: I1124 07:14:38.566044 4809 generic.go:334] "Generic (PLEG): container finished" podID="1286d658-6187-4509-9e24-1e0d6fbcab1e" containerID="0a3e1f620cd8d80a95f532ec6d9f8cb9f3a36e6385c2d51b423c6d336e475487" exitCode=143 Nov 24 07:14:38 crc kubenswrapper[4809]: I1124 07:14:38.566163 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1286d658-6187-4509-9e24-1e0d6fbcab1e","Type":"ContainerDied","Data":"0a3e1f620cd8d80a95f532ec6d9f8cb9f3a36e6385c2d51b423c6d336e475487"} Nov 24 07:14:38 crc kubenswrapper[4809]: I1124 07:14:38.568623 4809 generic.go:334] "Generic (PLEG): container finished" podID="02e88e29-10b0-42fb-96d3-f562aa48f361" containerID="b95add818e2dacbe46ab6681b06ab893d4b986180b8af6f266e102d9e3a57267" exitCode=143 Nov 24 07:14:38 crc kubenswrapper[4809]: I1124 07:14:38.568700 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"02e88e29-10b0-42fb-96d3-f562aa48f361","Type":"ContainerDied","Data":"b95add818e2dacbe46ab6681b06ab893d4b986180b8af6f266e102d9e3a57267"} Nov 24 07:14:40 crc kubenswrapper[4809]: I1124 07:14:40.850901 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="1286d658-6187-4509-9e24-1e0d6fbcab1e" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": read tcp 10.217.0.2:54118->10.217.0.194:8775: read: connection reset by peer" Nov 24 07:14:40 crc kubenswrapper[4809]: I1124 07:14:40.851154 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="1286d658-6187-4509-9e24-1e0d6fbcab1e" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": read tcp 10.217.0.2:54102->10.217.0.194:8775: read: connection reset by peer" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.281843 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.332270 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02e88e29-10b0-42fb-96d3-f562aa48f361-logs\") pod \"02e88e29-10b0-42fb-96d3-f562aa48f361\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.332342 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9z5f\" (UniqueName: \"kubernetes.io/projected/02e88e29-10b0-42fb-96d3-f562aa48f361-kube-api-access-c9z5f\") pod \"02e88e29-10b0-42fb-96d3-f562aa48f361\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.332370 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-public-tls-certs\") pod \"02e88e29-10b0-42fb-96d3-f562aa48f361\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.332411 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-config-data\") pod \"02e88e29-10b0-42fb-96d3-f562aa48f361\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.332499 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-combined-ca-bundle\") pod \"02e88e29-10b0-42fb-96d3-f562aa48f361\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.332607 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-internal-tls-certs\") pod \"02e88e29-10b0-42fb-96d3-f562aa48f361\" (UID: \"02e88e29-10b0-42fb-96d3-f562aa48f361\") " Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.333260 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02e88e29-10b0-42fb-96d3-f562aa48f361-logs" (OuterVolumeSpecName: "logs") pod "02e88e29-10b0-42fb-96d3-f562aa48f361" (UID: "02e88e29-10b0-42fb-96d3-f562aa48f361"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.333675 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02e88e29-10b0-42fb-96d3-f562aa48f361-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.338367 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02e88e29-10b0-42fb-96d3-f562aa48f361-kube-api-access-c9z5f" (OuterVolumeSpecName: "kube-api-access-c9z5f") pod "02e88e29-10b0-42fb-96d3-f562aa48f361" (UID: "02e88e29-10b0-42fb-96d3-f562aa48f361"). InnerVolumeSpecName "kube-api-access-c9z5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.346274 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.370583 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-config-data" (OuterVolumeSpecName: "config-data") pod "02e88e29-10b0-42fb-96d3-f562aa48f361" (UID: "02e88e29-10b0-42fb-96d3-f562aa48f361"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.396104 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02e88e29-10b0-42fb-96d3-f562aa48f361" (UID: "02e88e29-10b0-42fb-96d3-f562aa48f361"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.402826 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "02e88e29-10b0-42fb-96d3-f562aa48f361" (UID: "02e88e29-10b0-42fb-96d3-f562aa48f361"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.405294 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "02e88e29-10b0-42fb-96d3-f562aa48f361" (UID: "02e88e29-10b0-42fb-96d3-f562aa48f361"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.434739 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1286d658-6187-4509-9e24-1e0d6fbcab1e-combined-ca-bundle\") pod \"1286d658-6187-4509-9e24-1e0d6fbcab1e\" (UID: \"1286d658-6187-4509-9e24-1e0d6fbcab1e\") " Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.434884 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1286d658-6187-4509-9e24-1e0d6fbcab1e-logs\") pod \"1286d658-6187-4509-9e24-1e0d6fbcab1e\" (UID: \"1286d658-6187-4509-9e24-1e0d6fbcab1e\") " Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.434927 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztbps\" (UniqueName: \"kubernetes.io/projected/1286d658-6187-4509-9e24-1e0d6fbcab1e-kube-api-access-ztbps\") pod \"1286d658-6187-4509-9e24-1e0d6fbcab1e\" (UID: \"1286d658-6187-4509-9e24-1e0d6fbcab1e\") " Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.435043 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1286d658-6187-4509-9e24-1e0d6fbcab1e-nova-metadata-tls-certs\") pod \"1286d658-6187-4509-9e24-1e0d6fbcab1e\" (UID: \"1286d658-6187-4509-9e24-1e0d6fbcab1e\") " Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.435170 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1286d658-6187-4509-9e24-1e0d6fbcab1e-config-data\") pod \"1286d658-6187-4509-9e24-1e0d6fbcab1e\" (UID: \"1286d658-6187-4509-9e24-1e0d6fbcab1e\") " Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.435415 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1286d658-6187-4509-9e24-1e0d6fbcab1e-logs" (OuterVolumeSpecName: "logs") pod "1286d658-6187-4509-9e24-1e0d6fbcab1e" (UID: "1286d658-6187-4509-9e24-1e0d6fbcab1e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.436061 4809 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.436080 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1286d658-6187-4509-9e24-1e0d6fbcab1e-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.436094 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9z5f\" (UniqueName: \"kubernetes.io/projected/02e88e29-10b0-42fb-96d3-f562aa48f361-kube-api-access-c9z5f\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.436106 4809 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.436120 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.436132 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02e88e29-10b0-42fb-96d3-f562aa48f361-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.440028 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1286d658-6187-4509-9e24-1e0d6fbcab1e-kube-api-access-ztbps" (OuterVolumeSpecName: "kube-api-access-ztbps") pod "1286d658-6187-4509-9e24-1e0d6fbcab1e" (UID: "1286d658-6187-4509-9e24-1e0d6fbcab1e"). InnerVolumeSpecName "kube-api-access-ztbps". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.473441 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1286d658-6187-4509-9e24-1e0d6fbcab1e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1286d658-6187-4509-9e24-1e0d6fbcab1e" (UID: "1286d658-6187-4509-9e24-1e0d6fbcab1e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.474026 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1286d658-6187-4509-9e24-1e0d6fbcab1e-config-data" (OuterVolumeSpecName: "config-data") pod "1286d658-6187-4509-9e24-1e0d6fbcab1e" (UID: "1286d658-6187-4509-9e24-1e0d6fbcab1e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.514919 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1286d658-6187-4509-9e24-1e0d6fbcab1e-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "1286d658-6187-4509-9e24-1e0d6fbcab1e" (UID: "1286d658-6187-4509-9e24-1e0d6fbcab1e"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.538347 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1286d658-6187-4509-9e24-1e0d6fbcab1e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.538380 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1286d658-6187-4509-9e24-1e0d6fbcab1e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.538395 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztbps\" (UniqueName: \"kubernetes.io/projected/1286d658-6187-4509-9e24-1e0d6fbcab1e-kube-api-access-ztbps\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.538406 4809 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1286d658-6187-4509-9e24-1e0d6fbcab1e-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.600472 4809 generic.go:334] "Generic (PLEG): container finished" podID="1286d658-6187-4509-9e24-1e0d6fbcab1e" containerID="8252b5e689178dcecb351ad3aeeb9896ac6dbaa4ca0ff97c73985ad51a08b07b" exitCode=0 Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.600529 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1286d658-6187-4509-9e24-1e0d6fbcab1e","Type":"ContainerDied","Data":"8252b5e689178dcecb351ad3aeeb9896ac6dbaa4ca0ff97c73985ad51a08b07b"} Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.600558 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1286d658-6187-4509-9e24-1e0d6fbcab1e","Type":"ContainerDied","Data":"9e45992ea7bcf9d34a0895b715515999f4e06ab76942080ec1f9657fb87776c4"} Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.600573 4809 scope.go:117] "RemoveContainer" containerID="8252b5e689178dcecb351ad3aeeb9896ac6dbaa4ca0ff97c73985ad51a08b07b" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.600678 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.604818 4809 generic.go:334] "Generic (PLEG): container finished" podID="02e88e29-10b0-42fb-96d3-f562aa48f361" containerID="c44af3db3c24c41298e54d466fee3a1916bde42927dd20216ef9ac046f32783e" exitCode=0 Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.604849 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"02e88e29-10b0-42fb-96d3-f562aa48f361","Type":"ContainerDied","Data":"c44af3db3c24c41298e54d466fee3a1916bde42927dd20216ef9ac046f32783e"} Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.604872 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"02e88e29-10b0-42fb-96d3-f562aa48f361","Type":"ContainerDied","Data":"13a40a4f800906df01f104c13146716ecc1e7c8570ebadc2b2043b51d024e664"} Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.604918 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.627652 4809 scope.go:117] "RemoveContainer" containerID="0a3e1f620cd8d80a95f532ec6d9f8cb9f3a36e6385c2d51b423c6d336e475487" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.638166 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.646464 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.663657 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.675923 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 07:14:41 crc kubenswrapper[4809]: E1124 07:14:41.676523 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02e88e29-10b0-42fb-96d3-f562aa48f361" containerName="nova-api-api" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.676550 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="02e88e29-10b0-42fb-96d3-f562aa48f361" containerName="nova-api-api" Nov 24 07:14:41 crc kubenswrapper[4809]: E1124 07:14:41.676568 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1286d658-6187-4509-9e24-1e0d6fbcab1e" containerName="nova-metadata-metadata" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.676577 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1286d658-6187-4509-9e24-1e0d6fbcab1e" containerName="nova-metadata-metadata" Nov 24 07:14:41 crc kubenswrapper[4809]: E1124 07:14:41.676598 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1286d658-6187-4509-9e24-1e0d6fbcab1e" containerName="nova-metadata-log" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.676605 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1286d658-6187-4509-9e24-1e0d6fbcab1e" containerName="nova-metadata-log" Nov 24 07:14:41 crc kubenswrapper[4809]: E1124 07:14:41.676615 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02e88e29-10b0-42fb-96d3-f562aa48f361" containerName="nova-api-log" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.676622 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="02e88e29-10b0-42fb-96d3-f562aa48f361" containerName="nova-api-log" Nov 24 07:14:41 crc kubenswrapper[4809]: E1124 07:14:41.676655 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97266dae-027e-469f-8934-298e38fbde54" containerName="nova-manage" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.676664 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="97266dae-027e-469f-8934-298e38fbde54" containerName="nova-manage" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.676878 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1286d658-6187-4509-9e24-1e0d6fbcab1e" containerName="nova-metadata-log" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.676904 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="02e88e29-10b0-42fb-96d3-f562aa48f361" containerName="nova-api-api" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.676917 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="97266dae-027e-469f-8934-298e38fbde54" containerName="nova-manage" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.676935 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1286d658-6187-4509-9e24-1e0d6fbcab1e" containerName="nova-metadata-metadata" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.676943 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="02e88e29-10b0-42fb-96d3-f562aa48f361" containerName="nova-api-log" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.678166 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.683058 4809 scope.go:117] "RemoveContainer" containerID="8252b5e689178dcecb351ad3aeeb9896ac6dbaa4ca0ff97c73985ad51a08b07b" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.683545 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.683765 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 07:14:41 crc kubenswrapper[4809]: E1124 07:14:41.683815 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8252b5e689178dcecb351ad3aeeb9896ac6dbaa4ca0ff97c73985ad51a08b07b\": container with ID starting with 8252b5e689178dcecb351ad3aeeb9896ac6dbaa4ca0ff97c73985ad51a08b07b not found: ID does not exist" containerID="8252b5e689178dcecb351ad3aeeb9896ac6dbaa4ca0ff97c73985ad51a08b07b" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.683847 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8252b5e689178dcecb351ad3aeeb9896ac6dbaa4ca0ff97c73985ad51a08b07b"} err="failed to get container status \"8252b5e689178dcecb351ad3aeeb9896ac6dbaa4ca0ff97c73985ad51a08b07b\": rpc error: code = NotFound desc = could not find container \"8252b5e689178dcecb351ad3aeeb9896ac6dbaa4ca0ff97c73985ad51a08b07b\": container with ID starting with 8252b5e689178dcecb351ad3aeeb9896ac6dbaa4ca0ff97c73985ad51a08b07b not found: ID does not exist" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.683869 4809 scope.go:117] "RemoveContainer" containerID="0a3e1f620cd8d80a95f532ec6d9f8cb9f3a36e6385c2d51b423c6d336e475487" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.683851 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 07:14:41 crc kubenswrapper[4809]: E1124 07:14:41.687921 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a3e1f620cd8d80a95f532ec6d9f8cb9f3a36e6385c2d51b423c6d336e475487\": container with ID starting with 0a3e1f620cd8d80a95f532ec6d9f8cb9f3a36e6385c2d51b423c6d336e475487 not found: ID does not exist" containerID="0a3e1f620cd8d80a95f532ec6d9f8cb9f3a36e6385c2d51b423c6d336e475487" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.688803 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a3e1f620cd8d80a95f532ec6d9f8cb9f3a36e6385c2d51b423c6d336e475487"} err="failed to get container status \"0a3e1f620cd8d80a95f532ec6d9f8cb9f3a36e6385c2d51b423c6d336e475487\": rpc error: code = NotFound desc = could not find container \"0a3e1f620cd8d80a95f532ec6d9f8cb9f3a36e6385c2d51b423c6d336e475487\": container with ID starting with 0a3e1f620cd8d80a95f532ec6d9f8cb9f3a36e6385c2d51b423c6d336e475487 not found: ID does not exist" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.688910 4809 scope.go:117] "RemoveContainer" containerID="c44af3db3c24c41298e54d466fee3a1916bde42927dd20216ef9ac046f32783e" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.704955 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.722405 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.728165 4809 scope.go:117] "RemoveContainer" containerID="b95add818e2dacbe46ab6681b06ab893d4b986180b8af6f266e102d9e3a57267" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.732906 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.734708 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.740162 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.740438 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.741437 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9b4155b-3ac1-4c8d-b768-e28485799816-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d9b4155b-3ac1-4c8d-b768-e28485799816\") " pod="openstack/nova-metadata-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.741488 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2z8cr\" (UniqueName: \"kubernetes.io/projected/d9b4155b-3ac1-4c8d-b768-e28485799816-kube-api-access-2z8cr\") pod \"nova-metadata-0\" (UID: \"d9b4155b-3ac1-4c8d-b768-e28485799816\") " pod="openstack/nova-metadata-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.741511 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e8e43ff-8f38-4981-a78c-41934ecc1210-config-data\") pod \"nova-api-0\" (UID: \"6e8e43ff-8f38-4981-a78c-41934ecc1210\") " pod="openstack/nova-api-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.741538 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e8e43ff-8f38-4981-a78c-41934ecc1210-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6e8e43ff-8f38-4981-a78c-41934ecc1210\") " pod="openstack/nova-api-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.741558 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9b4155b-3ac1-4c8d-b768-e28485799816-config-data\") pod \"nova-metadata-0\" (UID: \"d9b4155b-3ac1-4c8d-b768-e28485799816\") " pod="openstack/nova-metadata-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.741587 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b4155b-3ac1-4c8d-b768-e28485799816-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d9b4155b-3ac1-4c8d-b768-e28485799816\") " pod="openstack/nova-metadata-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.741697 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e8e43ff-8f38-4981-a78c-41934ecc1210-public-tls-certs\") pod \"nova-api-0\" (UID: \"6e8e43ff-8f38-4981-a78c-41934ecc1210\") " pod="openstack/nova-api-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.741754 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhvjz\" (UniqueName: \"kubernetes.io/projected/6e8e43ff-8f38-4981-a78c-41934ecc1210-kube-api-access-qhvjz\") pod \"nova-api-0\" (UID: \"6e8e43ff-8f38-4981-a78c-41934ecc1210\") " pod="openstack/nova-api-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.741770 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e8e43ff-8f38-4981-a78c-41934ecc1210-logs\") pod \"nova-api-0\" (UID: \"6e8e43ff-8f38-4981-a78c-41934ecc1210\") " pod="openstack/nova-api-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.741813 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e8e43ff-8f38-4981-a78c-41934ecc1210-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6e8e43ff-8f38-4981-a78c-41934ecc1210\") " pod="openstack/nova-api-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.741868 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9b4155b-3ac1-4c8d-b768-e28485799816-logs\") pod \"nova-metadata-0\" (UID: \"d9b4155b-3ac1-4c8d-b768-e28485799816\") " pod="openstack/nova-metadata-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.757734 4809 scope.go:117] "RemoveContainer" containerID="c44af3db3c24c41298e54d466fee3a1916bde42927dd20216ef9ac046f32783e" Nov 24 07:14:41 crc kubenswrapper[4809]: E1124 07:14:41.758191 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c44af3db3c24c41298e54d466fee3a1916bde42927dd20216ef9ac046f32783e\": container with ID starting with c44af3db3c24c41298e54d466fee3a1916bde42927dd20216ef9ac046f32783e not found: ID does not exist" containerID="c44af3db3c24c41298e54d466fee3a1916bde42927dd20216ef9ac046f32783e" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.758224 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c44af3db3c24c41298e54d466fee3a1916bde42927dd20216ef9ac046f32783e"} err="failed to get container status \"c44af3db3c24c41298e54d466fee3a1916bde42927dd20216ef9ac046f32783e\": rpc error: code = NotFound desc = could not find container \"c44af3db3c24c41298e54d466fee3a1916bde42927dd20216ef9ac046f32783e\": container with ID starting with c44af3db3c24c41298e54d466fee3a1916bde42927dd20216ef9ac046f32783e not found: ID does not exist" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.758248 4809 scope.go:117] "RemoveContainer" containerID="b95add818e2dacbe46ab6681b06ab893d4b986180b8af6f266e102d9e3a57267" Nov 24 07:14:41 crc kubenswrapper[4809]: E1124 07:14:41.758548 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b95add818e2dacbe46ab6681b06ab893d4b986180b8af6f266e102d9e3a57267\": container with ID starting with b95add818e2dacbe46ab6681b06ab893d4b986180b8af6f266e102d9e3a57267 not found: ID does not exist" containerID="b95add818e2dacbe46ab6681b06ab893d4b986180b8af6f266e102d9e3a57267" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.758597 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b95add818e2dacbe46ab6681b06ab893d4b986180b8af6f266e102d9e3a57267"} err="failed to get container status \"b95add818e2dacbe46ab6681b06ab893d4b986180b8af6f266e102d9e3a57267\": rpc error: code = NotFound desc = could not find container \"b95add818e2dacbe46ab6681b06ab893d4b986180b8af6f266e102d9e3a57267\": container with ID starting with b95add818e2dacbe46ab6681b06ab893d4b986180b8af6f266e102d9e3a57267 not found: ID does not exist" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.767076 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.843613 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e8e43ff-8f38-4981-a78c-41934ecc1210-public-tls-certs\") pod \"nova-api-0\" (UID: \"6e8e43ff-8f38-4981-a78c-41934ecc1210\") " pod="openstack/nova-api-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.843905 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhvjz\" (UniqueName: \"kubernetes.io/projected/6e8e43ff-8f38-4981-a78c-41934ecc1210-kube-api-access-qhvjz\") pod \"nova-api-0\" (UID: \"6e8e43ff-8f38-4981-a78c-41934ecc1210\") " pod="openstack/nova-api-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.844093 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e8e43ff-8f38-4981-a78c-41934ecc1210-logs\") pod \"nova-api-0\" (UID: \"6e8e43ff-8f38-4981-a78c-41934ecc1210\") " pod="openstack/nova-api-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.844505 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e8e43ff-8f38-4981-a78c-41934ecc1210-logs\") pod \"nova-api-0\" (UID: \"6e8e43ff-8f38-4981-a78c-41934ecc1210\") " pod="openstack/nova-api-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.844737 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e8e43ff-8f38-4981-a78c-41934ecc1210-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6e8e43ff-8f38-4981-a78c-41934ecc1210\") " pod="openstack/nova-api-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.844907 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9b4155b-3ac1-4c8d-b768-e28485799816-logs\") pod \"nova-metadata-0\" (UID: \"d9b4155b-3ac1-4c8d-b768-e28485799816\") " pod="openstack/nova-metadata-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.845035 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9b4155b-3ac1-4c8d-b768-e28485799816-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d9b4155b-3ac1-4c8d-b768-e28485799816\") " pod="openstack/nova-metadata-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.845177 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2z8cr\" (UniqueName: \"kubernetes.io/projected/d9b4155b-3ac1-4c8d-b768-e28485799816-kube-api-access-2z8cr\") pod \"nova-metadata-0\" (UID: \"d9b4155b-3ac1-4c8d-b768-e28485799816\") " pod="openstack/nova-metadata-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.845281 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e8e43ff-8f38-4981-a78c-41934ecc1210-config-data\") pod \"nova-api-0\" (UID: \"6e8e43ff-8f38-4981-a78c-41934ecc1210\") " pod="openstack/nova-api-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.845359 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9b4155b-3ac1-4c8d-b768-e28485799816-logs\") pod \"nova-metadata-0\" (UID: \"d9b4155b-3ac1-4c8d-b768-e28485799816\") " pod="openstack/nova-metadata-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.845466 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e8e43ff-8f38-4981-a78c-41934ecc1210-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6e8e43ff-8f38-4981-a78c-41934ecc1210\") " pod="openstack/nova-api-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.845562 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9b4155b-3ac1-4c8d-b768-e28485799816-config-data\") pod \"nova-metadata-0\" (UID: \"d9b4155b-3ac1-4c8d-b768-e28485799816\") " pod="openstack/nova-metadata-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.845685 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b4155b-3ac1-4c8d-b768-e28485799816-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d9b4155b-3ac1-4c8d-b768-e28485799816\") " pod="openstack/nova-metadata-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.847846 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e8e43ff-8f38-4981-a78c-41934ecc1210-public-tls-certs\") pod \"nova-api-0\" (UID: \"6e8e43ff-8f38-4981-a78c-41934ecc1210\") " pod="openstack/nova-api-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.848024 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9b4155b-3ac1-4c8d-b768-e28485799816-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d9b4155b-3ac1-4c8d-b768-e28485799816\") " pod="openstack/nova-metadata-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.848737 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e8e43ff-8f38-4981-a78c-41934ecc1210-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6e8e43ff-8f38-4981-a78c-41934ecc1210\") " pod="openstack/nova-api-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.849009 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b4155b-3ac1-4c8d-b768-e28485799816-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d9b4155b-3ac1-4c8d-b768-e28485799816\") " pod="openstack/nova-metadata-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.849655 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e8e43ff-8f38-4981-a78c-41934ecc1210-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6e8e43ff-8f38-4981-a78c-41934ecc1210\") " pod="openstack/nova-api-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.852231 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9b4155b-3ac1-4c8d-b768-e28485799816-config-data\") pod \"nova-metadata-0\" (UID: \"d9b4155b-3ac1-4c8d-b768-e28485799816\") " pod="openstack/nova-metadata-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.852930 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e8e43ff-8f38-4981-a78c-41934ecc1210-config-data\") pod \"nova-api-0\" (UID: \"6e8e43ff-8f38-4981-a78c-41934ecc1210\") " pod="openstack/nova-api-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.861205 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhvjz\" (UniqueName: \"kubernetes.io/projected/6e8e43ff-8f38-4981-a78c-41934ecc1210-kube-api-access-qhvjz\") pod \"nova-api-0\" (UID: \"6e8e43ff-8f38-4981-a78c-41934ecc1210\") " pod="openstack/nova-api-0" Nov 24 07:14:41 crc kubenswrapper[4809]: I1124 07:14:41.862846 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2z8cr\" (UniqueName: \"kubernetes.io/projected/d9b4155b-3ac1-4c8d-b768-e28485799816-kube-api-access-2z8cr\") pod \"nova-metadata-0\" (UID: \"d9b4155b-3ac1-4c8d-b768-e28485799816\") " pod="openstack/nova-metadata-0" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.011715 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.057933 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.399582 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.456671 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/111098ed-9d16-45e0-bb80-35f9383e327e-combined-ca-bundle\") pod \"111098ed-9d16-45e0-bb80-35f9383e327e\" (UID: \"111098ed-9d16-45e0-bb80-35f9383e327e\") " Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.456772 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cw4gs\" (UniqueName: \"kubernetes.io/projected/111098ed-9d16-45e0-bb80-35f9383e327e-kube-api-access-cw4gs\") pod \"111098ed-9d16-45e0-bb80-35f9383e327e\" (UID: \"111098ed-9d16-45e0-bb80-35f9383e327e\") " Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.456802 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/111098ed-9d16-45e0-bb80-35f9383e327e-config-data\") pod \"111098ed-9d16-45e0-bb80-35f9383e327e\" (UID: \"111098ed-9d16-45e0-bb80-35f9383e327e\") " Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.461628 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/111098ed-9d16-45e0-bb80-35f9383e327e-kube-api-access-cw4gs" (OuterVolumeSpecName: "kube-api-access-cw4gs") pod "111098ed-9d16-45e0-bb80-35f9383e327e" (UID: "111098ed-9d16-45e0-bb80-35f9383e327e"). InnerVolumeSpecName "kube-api-access-cw4gs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.483704 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/111098ed-9d16-45e0-bb80-35f9383e327e-config-data" (OuterVolumeSpecName: "config-data") pod "111098ed-9d16-45e0-bb80-35f9383e327e" (UID: "111098ed-9d16-45e0-bb80-35f9383e327e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.488780 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/111098ed-9d16-45e0-bb80-35f9383e327e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "111098ed-9d16-45e0-bb80-35f9383e327e" (UID: "111098ed-9d16-45e0-bb80-35f9383e327e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.548684 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.559008 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.560202 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/111098ed-9d16-45e0-bb80-35f9383e327e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.560237 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cw4gs\" (UniqueName: \"kubernetes.io/projected/111098ed-9d16-45e0-bb80-35f9383e327e-kube-api-access-cw4gs\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.560250 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/111098ed-9d16-45e0-bb80-35f9383e327e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:14:42 crc kubenswrapper[4809]: W1124 07:14:42.573509 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9b4155b_3ac1_4c8d_b768_e28485799816.slice/crio-3d489dab98c66ed7fafe9b801302e70fdf5545919b0ebcb2bd704cf196ed9096 WatchSource:0}: Error finding container 3d489dab98c66ed7fafe9b801302e70fdf5545919b0ebcb2bd704cf196ed9096: Status 404 returned error can't find the container with id 3d489dab98c66ed7fafe9b801302e70fdf5545919b0ebcb2bd704cf196ed9096 Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.616025 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d9b4155b-3ac1-4c8d-b768-e28485799816","Type":"ContainerStarted","Data":"3d489dab98c66ed7fafe9b801302e70fdf5545919b0ebcb2bd704cf196ed9096"} Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.621779 4809 generic.go:334] "Generic (PLEG): container finished" podID="111098ed-9d16-45e0-bb80-35f9383e327e" containerID="ffb9df9f2296363c66395fba7c7aa3efbc5800406624655d76c1096454276be2" exitCode=0 Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.621838 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.621845 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"111098ed-9d16-45e0-bb80-35f9383e327e","Type":"ContainerDied","Data":"ffb9df9f2296363c66395fba7c7aa3efbc5800406624655d76c1096454276be2"} Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.621867 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"111098ed-9d16-45e0-bb80-35f9383e327e","Type":"ContainerDied","Data":"276bdc3c5eb76ef9fadb7fc2fe909c352901ea24ae7bade45885aaa3a2991ef6"} Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.621896 4809 scope.go:117] "RemoveContainer" containerID="ffb9df9f2296363c66395fba7c7aa3efbc5800406624655d76c1096454276be2" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.623933 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e8e43ff-8f38-4981-a78c-41934ecc1210","Type":"ContainerStarted","Data":"e050525628c3e48dfa9abd324eb48ab4b1ce333f7026bd3fa1a1c9d2725f027b"} Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.660833 4809 scope.go:117] "RemoveContainer" containerID="ffb9df9f2296363c66395fba7c7aa3efbc5800406624655d76c1096454276be2" Nov 24 07:14:42 crc kubenswrapper[4809]: E1124 07:14:42.668951 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffb9df9f2296363c66395fba7c7aa3efbc5800406624655d76c1096454276be2\": container with ID starting with ffb9df9f2296363c66395fba7c7aa3efbc5800406624655d76c1096454276be2 not found: ID does not exist" containerID="ffb9df9f2296363c66395fba7c7aa3efbc5800406624655d76c1096454276be2" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.669006 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffb9df9f2296363c66395fba7c7aa3efbc5800406624655d76c1096454276be2"} err="failed to get container status \"ffb9df9f2296363c66395fba7c7aa3efbc5800406624655d76c1096454276be2\": rpc error: code = NotFound desc = could not find container \"ffb9df9f2296363c66395fba7c7aa3efbc5800406624655d76c1096454276be2\": container with ID starting with ffb9df9f2296363c66395fba7c7aa3efbc5800406624655d76c1096454276be2 not found: ID does not exist" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.679907 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.695886 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.722702 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:14:42 crc kubenswrapper[4809]: E1124 07:14:42.723449 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="111098ed-9d16-45e0-bb80-35f9383e327e" containerName="nova-scheduler-scheduler" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.723467 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="111098ed-9d16-45e0-bb80-35f9383e327e" containerName="nova-scheduler-scheduler" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.723844 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="111098ed-9d16-45e0-bb80-35f9383e327e" containerName="nova-scheduler-scheduler" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.725164 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.727751 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.743474 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.763008 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwhqk\" (UniqueName: \"kubernetes.io/projected/d4270af3-165e-49e5-b88e-1c07b1a3a950-kube-api-access-dwhqk\") pod \"nova-scheduler-0\" (UID: \"d4270af3-165e-49e5-b88e-1c07b1a3a950\") " pod="openstack/nova-scheduler-0" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.763193 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4270af3-165e-49e5-b88e-1c07b1a3a950-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d4270af3-165e-49e5-b88e-1c07b1a3a950\") " pod="openstack/nova-scheduler-0" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.763250 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4270af3-165e-49e5-b88e-1c07b1a3a950-config-data\") pod \"nova-scheduler-0\" (UID: \"d4270af3-165e-49e5-b88e-1c07b1a3a950\") " pod="openstack/nova-scheduler-0" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.865067 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwhqk\" (UniqueName: \"kubernetes.io/projected/d4270af3-165e-49e5-b88e-1c07b1a3a950-kube-api-access-dwhqk\") pod \"nova-scheduler-0\" (UID: \"d4270af3-165e-49e5-b88e-1c07b1a3a950\") " pod="openstack/nova-scheduler-0" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.865243 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4270af3-165e-49e5-b88e-1c07b1a3a950-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d4270af3-165e-49e5-b88e-1c07b1a3a950\") " pod="openstack/nova-scheduler-0" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.865278 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4270af3-165e-49e5-b88e-1c07b1a3a950-config-data\") pod \"nova-scheduler-0\" (UID: \"d4270af3-165e-49e5-b88e-1c07b1a3a950\") " pod="openstack/nova-scheduler-0" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.874071 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4270af3-165e-49e5-b88e-1c07b1a3a950-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d4270af3-165e-49e5-b88e-1c07b1a3a950\") " pod="openstack/nova-scheduler-0" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.878626 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4270af3-165e-49e5-b88e-1c07b1a3a950-config-data\") pod \"nova-scheduler-0\" (UID: \"d4270af3-165e-49e5-b88e-1c07b1a3a950\") " pod="openstack/nova-scheduler-0" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.883822 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwhqk\" (UniqueName: \"kubernetes.io/projected/d4270af3-165e-49e5-b88e-1c07b1a3a950-kube-api-access-dwhqk\") pod \"nova-scheduler-0\" (UID: \"d4270af3-165e-49e5-b88e-1c07b1a3a950\") " pod="openstack/nova-scheduler-0" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.902444 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02e88e29-10b0-42fb-96d3-f562aa48f361" path="/var/lib/kubelet/pods/02e88e29-10b0-42fb-96d3-f562aa48f361/volumes" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.903366 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="111098ed-9d16-45e0-bb80-35f9383e327e" path="/var/lib/kubelet/pods/111098ed-9d16-45e0-bb80-35f9383e327e/volumes" Nov 24 07:14:42 crc kubenswrapper[4809]: I1124 07:14:42.904054 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1286d658-6187-4509-9e24-1e0d6fbcab1e" path="/var/lib/kubelet/pods/1286d658-6187-4509-9e24-1e0d6fbcab1e/volumes" Nov 24 07:14:43 crc kubenswrapper[4809]: I1124 07:14:43.046037 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 07:14:43 crc kubenswrapper[4809]: I1124 07:14:43.509181 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:14:43 crc kubenswrapper[4809]: I1124 07:14:43.638701 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d4270af3-165e-49e5-b88e-1c07b1a3a950","Type":"ContainerStarted","Data":"b5d54be27adfecb48775dfa2dd4804462ab618923a94ee414656d98f388971b7"} Nov 24 07:14:43 crc kubenswrapper[4809]: I1124 07:14:43.641502 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d9b4155b-3ac1-4c8d-b768-e28485799816","Type":"ContainerStarted","Data":"c03d9dc5f765063e1cafddd33c0c1a3b4e56a2fcfa2e80193cf10e359161f277"} Nov 24 07:14:43 crc kubenswrapper[4809]: I1124 07:14:43.641569 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d9b4155b-3ac1-4c8d-b768-e28485799816","Type":"ContainerStarted","Data":"6e590a00e1eb3d44c1481e209957827e7a0de2c95b3b09f40dcd592dd017b3f0"} Nov 24 07:14:43 crc kubenswrapper[4809]: I1124 07:14:43.647676 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e8e43ff-8f38-4981-a78c-41934ecc1210","Type":"ContainerStarted","Data":"b869b97d3844655dee2743a12740057bef92e300d5b7c06180947e7c12f3df0d"} Nov 24 07:14:43 crc kubenswrapper[4809]: I1124 07:14:43.647746 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e8e43ff-8f38-4981-a78c-41934ecc1210","Type":"ContainerStarted","Data":"5fe5d84c9208269275cd5f651f7c39ef8e17c69ff9b4d0776439fcd43055f00a"} Nov 24 07:14:43 crc kubenswrapper[4809]: I1124 07:14:43.676656 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.676628782 podStartE2EDuration="2.676628782s" podCreationTimestamp="2025-11-24 07:14:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:14:43.660924825 +0000 UTC m=+1183.561516570" watchObservedRunningTime="2025-11-24 07:14:43.676628782 +0000 UTC m=+1183.577220527" Nov 24 07:14:43 crc kubenswrapper[4809]: I1124 07:14:43.699262 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.699243991 podStartE2EDuration="2.699243991s" podCreationTimestamp="2025-11-24 07:14:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:14:43.694446584 +0000 UTC m=+1183.595038299" watchObservedRunningTime="2025-11-24 07:14:43.699243991 +0000 UTC m=+1183.599835706" Nov 24 07:14:44 crc kubenswrapper[4809]: I1124 07:14:44.674330 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d4270af3-165e-49e5-b88e-1c07b1a3a950","Type":"ContainerStarted","Data":"0fb027dee74271367ea7ad98193c00122781bf3eb37a9024574d64101f197d30"} Nov 24 07:14:44 crc kubenswrapper[4809]: I1124 07:14:44.702489 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.702466838 podStartE2EDuration="2.702466838s" podCreationTimestamp="2025-11-24 07:14:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:14:44.692236357 +0000 UTC m=+1184.592828062" watchObservedRunningTime="2025-11-24 07:14:44.702466838 +0000 UTC m=+1184.603058543" Nov 24 07:14:47 crc kubenswrapper[4809]: I1124 07:14:47.058223 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 07:14:47 crc kubenswrapper[4809]: I1124 07:14:47.058504 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 07:14:48 crc kubenswrapper[4809]: I1124 07:14:48.046464 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 07:14:52 crc kubenswrapper[4809]: I1124 07:14:52.012166 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 07:14:52 crc kubenswrapper[4809]: I1124 07:14:52.012869 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 07:14:52 crc kubenswrapper[4809]: I1124 07:14:52.058949 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 07:14:52 crc kubenswrapper[4809]: I1124 07:14:52.059064 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 07:14:53 crc kubenswrapper[4809]: I1124 07:14:53.032204 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6e8e43ff-8f38-4981-a78c-41934ecc1210" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 07:14:53 crc kubenswrapper[4809]: I1124 07:14:53.032233 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6e8e43ff-8f38-4981-a78c-41934ecc1210" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 07:14:53 crc kubenswrapper[4809]: I1124 07:14:53.047188 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 07:14:53 crc kubenswrapper[4809]: I1124 07:14:53.075220 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 07:14:53 crc kubenswrapper[4809]: I1124 07:14:53.076224 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d9b4155b-3ac1-4c8d-b768-e28485799816" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 07:14:53 crc kubenswrapper[4809]: I1124 07:14:53.076235 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d9b4155b-3ac1-4c8d-b768-e28485799816" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 07:14:53 crc kubenswrapper[4809]: I1124 07:14:53.818325 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 07:14:58 crc kubenswrapper[4809]: I1124 07:14:58.870122 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 07:15:00 crc kubenswrapper[4809]: I1124 07:15:00.141519 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399475-h4sk5"] Nov 24 07:15:00 crc kubenswrapper[4809]: I1124 07:15:00.142909 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-h4sk5" Nov 24 07:15:00 crc kubenswrapper[4809]: I1124 07:15:00.146549 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 07:15:00 crc kubenswrapper[4809]: I1124 07:15:00.146641 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 07:15:00 crc kubenswrapper[4809]: I1124 07:15:00.154500 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399475-h4sk5"] Nov 24 07:15:00 crc kubenswrapper[4809]: I1124 07:15:00.315242 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b19ae12c-c3c5-4082-be4f-ffde1d1f56f0-secret-volume\") pod \"collect-profiles-29399475-h4sk5\" (UID: \"b19ae12c-c3c5-4082-be4f-ffde1d1f56f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-h4sk5" Nov 24 07:15:00 crc kubenswrapper[4809]: I1124 07:15:00.315414 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfqj9\" (UniqueName: \"kubernetes.io/projected/b19ae12c-c3c5-4082-be4f-ffde1d1f56f0-kube-api-access-wfqj9\") pod \"collect-profiles-29399475-h4sk5\" (UID: \"b19ae12c-c3c5-4082-be4f-ffde1d1f56f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-h4sk5" Nov 24 07:15:00 crc kubenswrapper[4809]: I1124 07:15:00.315465 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b19ae12c-c3c5-4082-be4f-ffde1d1f56f0-config-volume\") pod \"collect-profiles-29399475-h4sk5\" (UID: \"b19ae12c-c3c5-4082-be4f-ffde1d1f56f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-h4sk5" Nov 24 07:15:00 crc kubenswrapper[4809]: I1124 07:15:00.417529 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b19ae12c-c3c5-4082-be4f-ffde1d1f56f0-config-volume\") pod \"collect-profiles-29399475-h4sk5\" (UID: \"b19ae12c-c3c5-4082-be4f-ffde1d1f56f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-h4sk5" Nov 24 07:15:00 crc kubenswrapper[4809]: I1124 07:15:00.417647 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b19ae12c-c3c5-4082-be4f-ffde1d1f56f0-secret-volume\") pod \"collect-profiles-29399475-h4sk5\" (UID: \"b19ae12c-c3c5-4082-be4f-ffde1d1f56f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-h4sk5" Nov 24 07:15:00 crc kubenswrapper[4809]: I1124 07:15:00.417753 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfqj9\" (UniqueName: \"kubernetes.io/projected/b19ae12c-c3c5-4082-be4f-ffde1d1f56f0-kube-api-access-wfqj9\") pod \"collect-profiles-29399475-h4sk5\" (UID: \"b19ae12c-c3c5-4082-be4f-ffde1d1f56f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-h4sk5" Nov 24 07:15:00 crc kubenswrapper[4809]: I1124 07:15:00.418861 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b19ae12c-c3c5-4082-be4f-ffde1d1f56f0-config-volume\") pod \"collect-profiles-29399475-h4sk5\" (UID: \"b19ae12c-c3c5-4082-be4f-ffde1d1f56f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-h4sk5" Nov 24 07:15:00 crc kubenswrapper[4809]: I1124 07:15:00.428934 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b19ae12c-c3c5-4082-be4f-ffde1d1f56f0-secret-volume\") pod \"collect-profiles-29399475-h4sk5\" (UID: \"b19ae12c-c3c5-4082-be4f-ffde1d1f56f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-h4sk5" Nov 24 07:15:00 crc kubenswrapper[4809]: I1124 07:15:00.442408 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfqj9\" (UniqueName: \"kubernetes.io/projected/b19ae12c-c3c5-4082-be4f-ffde1d1f56f0-kube-api-access-wfqj9\") pod \"collect-profiles-29399475-h4sk5\" (UID: \"b19ae12c-c3c5-4082-be4f-ffde1d1f56f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-h4sk5" Nov 24 07:15:00 crc kubenswrapper[4809]: I1124 07:15:00.471292 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-h4sk5" Nov 24 07:15:00 crc kubenswrapper[4809]: W1124 07:15:00.936445 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb19ae12c_c3c5_4082_be4f_ffde1d1f56f0.slice/crio-33593268afdfcb246afbfbe0c476dfbd962a98efbb54df04d45e0169977a52a6 WatchSource:0}: Error finding container 33593268afdfcb246afbfbe0c476dfbd962a98efbb54df04d45e0169977a52a6: Status 404 returned error can't find the container with id 33593268afdfcb246afbfbe0c476dfbd962a98efbb54df04d45e0169977a52a6 Nov 24 07:15:00 crc kubenswrapper[4809]: I1124 07:15:00.941102 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399475-h4sk5"] Nov 24 07:15:01 crc kubenswrapper[4809]: I1124 07:15:01.881423 4809 generic.go:334] "Generic (PLEG): container finished" podID="b19ae12c-c3c5-4082-be4f-ffde1d1f56f0" containerID="093604f5c39be5a54b49888cf4e0f869d9d6902fc20cb31eb8fccf902e3574d2" exitCode=0 Nov 24 07:15:01 crc kubenswrapper[4809]: I1124 07:15:01.881919 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-h4sk5" event={"ID":"b19ae12c-c3c5-4082-be4f-ffde1d1f56f0","Type":"ContainerDied","Data":"093604f5c39be5a54b49888cf4e0f869d9d6902fc20cb31eb8fccf902e3574d2"} Nov 24 07:15:01 crc kubenswrapper[4809]: I1124 07:15:01.881956 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-h4sk5" event={"ID":"b19ae12c-c3c5-4082-be4f-ffde1d1f56f0","Type":"ContainerStarted","Data":"33593268afdfcb246afbfbe0c476dfbd962a98efbb54df04d45e0169977a52a6"} Nov 24 07:15:02 crc kubenswrapper[4809]: I1124 07:15:02.019148 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 07:15:02 crc kubenswrapper[4809]: I1124 07:15:02.019759 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 07:15:02 crc kubenswrapper[4809]: I1124 07:15:02.020348 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 07:15:02 crc kubenswrapper[4809]: I1124 07:15:02.027045 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 07:15:02 crc kubenswrapper[4809]: I1124 07:15:02.064248 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 07:15:02 crc kubenswrapper[4809]: I1124 07:15:02.064327 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 07:15:02 crc kubenswrapper[4809]: I1124 07:15:02.068986 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 07:15:02 crc kubenswrapper[4809]: I1124 07:15:02.069044 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 07:15:02 crc kubenswrapper[4809]: I1124 07:15:02.906001 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 07:15:02 crc kubenswrapper[4809]: I1124 07:15:02.906307 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 07:15:03 crc kubenswrapper[4809]: I1124 07:15:03.297712 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-h4sk5" Nov 24 07:15:03 crc kubenswrapper[4809]: I1124 07:15:03.480768 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b19ae12c-c3c5-4082-be4f-ffde1d1f56f0-config-volume\") pod \"b19ae12c-c3c5-4082-be4f-ffde1d1f56f0\" (UID: \"b19ae12c-c3c5-4082-be4f-ffde1d1f56f0\") " Nov 24 07:15:03 crc kubenswrapper[4809]: I1124 07:15:03.480960 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfqj9\" (UniqueName: \"kubernetes.io/projected/b19ae12c-c3c5-4082-be4f-ffde1d1f56f0-kube-api-access-wfqj9\") pod \"b19ae12c-c3c5-4082-be4f-ffde1d1f56f0\" (UID: \"b19ae12c-c3c5-4082-be4f-ffde1d1f56f0\") " Nov 24 07:15:03 crc kubenswrapper[4809]: I1124 07:15:03.481021 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b19ae12c-c3c5-4082-be4f-ffde1d1f56f0-secret-volume\") pod \"b19ae12c-c3c5-4082-be4f-ffde1d1f56f0\" (UID: \"b19ae12c-c3c5-4082-be4f-ffde1d1f56f0\") " Nov 24 07:15:03 crc kubenswrapper[4809]: I1124 07:15:03.481564 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b19ae12c-c3c5-4082-be4f-ffde1d1f56f0-config-volume" (OuterVolumeSpecName: "config-volume") pod "b19ae12c-c3c5-4082-be4f-ffde1d1f56f0" (UID: "b19ae12c-c3c5-4082-be4f-ffde1d1f56f0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:15:03 crc kubenswrapper[4809]: I1124 07:15:03.497372 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b19ae12c-c3c5-4082-be4f-ffde1d1f56f0-kube-api-access-wfqj9" (OuterVolumeSpecName: "kube-api-access-wfqj9") pod "b19ae12c-c3c5-4082-be4f-ffde1d1f56f0" (UID: "b19ae12c-c3c5-4082-be4f-ffde1d1f56f0"). InnerVolumeSpecName "kube-api-access-wfqj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:15:03 crc kubenswrapper[4809]: I1124 07:15:03.499085 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b19ae12c-c3c5-4082-be4f-ffde1d1f56f0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b19ae12c-c3c5-4082-be4f-ffde1d1f56f0" (UID: "b19ae12c-c3c5-4082-be4f-ffde1d1f56f0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:15:03 crc kubenswrapper[4809]: I1124 07:15:03.583311 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfqj9\" (UniqueName: \"kubernetes.io/projected/b19ae12c-c3c5-4082-be4f-ffde1d1f56f0-kube-api-access-wfqj9\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:03 crc kubenswrapper[4809]: I1124 07:15:03.583366 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b19ae12c-c3c5-4082-be4f-ffde1d1f56f0-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:03 crc kubenswrapper[4809]: I1124 07:15:03.583379 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b19ae12c-c3c5-4082-be4f-ffde1d1f56f0-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:03 crc kubenswrapper[4809]: I1124 07:15:03.904215 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-h4sk5" Nov 24 07:15:03 crc kubenswrapper[4809]: I1124 07:15:03.904301 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-h4sk5" event={"ID":"b19ae12c-c3c5-4082-be4f-ffde1d1f56f0","Type":"ContainerDied","Data":"33593268afdfcb246afbfbe0c476dfbd962a98efbb54df04d45e0169977a52a6"} Nov 24 07:15:03 crc kubenswrapper[4809]: I1124 07:15:03.904343 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33593268afdfcb246afbfbe0c476dfbd962a98efbb54df04d45e0169977a52a6" Nov 24 07:15:10 crc kubenswrapper[4809]: I1124 07:15:10.906950 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:15:11 crc kubenswrapper[4809]: I1124 07:15:11.913121 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:15:14 crc kubenswrapper[4809]: I1124 07:15:14.708523 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="7d0a537c-b1f0-4908-a604-424ea6a58438" containerName="rabbitmq" containerID="cri-o://20a05ddef049ee32e97c296b2435378bf343ba8c54800ac573d5e1caa2bfa200" gracePeriod=604797 Nov 24 07:15:15 crc kubenswrapper[4809]: I1124 07:15:15.947069 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="f91f6ab4-0c7c-4002-aad9-dcd408c9000e" containerName="rabbitmq" containerID="cri-o://d6cfe70d34456cc68d198ac9f1cb919fdebac98b1adbc19e4e8accb2affa5c81" gracePeriod=604796 Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.144568 4809 generic.go:334] "Generic (PLEG): container finished" podID="7d0a537c-b1f0-4908-a604-424ea6a58438" containerID="20a05ddef049ee32e97c296b2435378bf343ba8c54800ac573d5e1caa2bfa200" exitCode=0 Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.145074 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7d0a537c-b1f0-4908-a604-424ea6a58438","Type":"ContainerDied","Data":"20a05ddef049ee32e97c296b2435378bf343ba8c54800ac573d5e1caa2bfa200"} Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.370452 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.464577 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7cjz\" (UniqueName: \"kubernetes.io/projected/7d0a537c-b1f0-4908-a604-424ea6a58438-kube-api-access-w7cjz\") pod \"7d0a537c-b1f0-4908-a604-424ea6a58438\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.464639 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-erlang-cookie\") pod \"7d0a537c-b1f0-4908-a604-424ea6a58438\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.464676 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7d0a537c-b1f0-4908-a604-424ea6a58438-plugins-conf\") pod \"7d0a537c-b1f0-4908-a604-424ea6a58438\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.464710 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7d0a537c-b1f0-4908-a604-424ea6a58438-erlang-cookie-secret\") pod \"7d0a537c-b1f0-4908-a604-424ea6a58438\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.464760 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"7d0a537c-b1f0-4908-a604-424ea6a58438\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.464789 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7d0a537c-b1f0-4908-a604-424ea6a58438-server-conf\") pod \"7d0a537c-b1f0-4908-a604-424ea6a58438\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.464845 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-plugins\") pod \"7d0a537c-b1f0-4908-a604-424ea6a58438\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.464897 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-tls\") pod \"7d0a537c-b1f0-4908-a604-424ea6a58438\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.464942 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-confd\") pod \"7d0a537c-b1f0-4908-a604-424ea6a58438\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.465002 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7d0a537c-b1f0-4908-a604-424ea6a58438-config-data\") pod \"7d0a537c-b1f0-4908-a604-424ea6a58438\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.465041 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7d0a537c-b1f0-4908-a604-424ea6a58438-pod-info\") pod \"7d0a537c-b1f0-4908-a604-424ea6a58438\" (UID: \"7d0a537c-b1f0-4908-a604-424ea6a58438\") " Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.466139 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "7d0a537c-b1f0-4908-a604-424ea6a58438" (UID: "7d0a537c-b1f0-4908-a604-424ea6a58438"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.466249 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "7d0a537c-b1f0-4908-a604-424ea6a58438" (UID: "7d0a537c-b1f0-4908-a604-424ea6a58438"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.467214 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d0a537c-b1f0-4908-a604-424ea6a58438-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "7d0a537c-b1f0-4908-a604-424ea6a58438" (UID: "7d0a537c-b1f0-4908-a604-424ea6a58438"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.472392 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d0a537c-b1f0-4908-a604-424ea6a58438-kube-api-access-w7cjz" (OuterVolumeSpecName: "kube-api-access-w7cjz") pod "7d0a537c-b1f0-4908-a604-424ea6a58438" (UID: "7d0a537c-b1f0-4908-a604-424ea6a58438"). InnerVolumeSpecName "kube-api-access-w7cjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.474305 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d0a537c-b1f0-4908-a604-424ea6a58438-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "7d0a537c-b1f0-4908-a604-424ea6a58438" (UID: "7d0a537c-b1f0-4908-a604-424ea6a58438"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.475818 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "7d0a537c-b1f0-4908-a604-424ea6a58438" (UID: "7d0a537c-b1f0-4908-a604-424ea6a58438"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.482679 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/7d0a537c-b1f0-4908-a604-424ea6a58438-pod-info" (OuterVolumeSpecName: "pod-info") pod "7d0a537c-b1f0-4908-a604-424ea6a58438" (UID: "7d0a537c-b1f0-4908-a604-424ea6a58438"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.493467 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "persistence") pod "7d0a537c-b1f0-4908-a604-424ea6a58438" (UID: "7d0a537c-b1f0-4908-a604-424ea6a58438"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.507593 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d0a537c-b1f0-4908-a604-424ea6a58438-config-data" (OuterVolumeSpecName: "config-data") pod "7d0a537c-b1f0-4908-a604-424ea6a58438" (UID: "7d0a537c-b1f0-4908-a604-424ea6a58438"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.569796 4809 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7d0a537c-b1f0-4908-a604-424ea6a58438-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.570412 4809 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7d0a537c-b1f0-4908-a604-424ea6a58438-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.570913 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.571071 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.571177 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.571256 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7d0a537c-b1f0-4908-a604-424ea6a58438-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.571366 4809 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7d0a537c-b1f0-4908-a604-424ea6a58438-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.571464 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7cjz\" (UniqueName: \"kubernetes.io/projected/7d0a537c-b1f0-4908-a604-424ea6a58438-kube-api-access-w7cjz\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.571551 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.607240 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d0a537c-b1f0-4908-a604-424ea6a58438-server-conf" (OuterVolumeSpecName: "server-conf") pod "7d0a537c-b1f0-4908-a604-424ea6a58438" (UID: "7d0a537c-b1f0-4908-a604-424ea6a58438"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.618616 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.637109 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "7d0a537c-b1f0-4908-a604-424ea6a58438" (UID: "7d0a537c-b1f0-4908-a604-424ea6a58438"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.672656 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.672690 4809 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7d0a537c-b1f0-4908-a604-424ea6a58438-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:21 crc kubenswrapper[4809]: I1124 07:15:21.672707 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7d0a537c-b1f0-4908-a604-424ea6a58438-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.155159 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7d0a537c-b1f0-4908-a604-424ea6a58438","Type":"ContainerDied","Data":"64b042e315d4bc4990612df074423d1bf871c1b70bca6735d37406e07036dcb3"} Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.155222 4809 scope.go:117] "RemoveContainer" containerID="20a05ddef049ee32e97c296b2435378bf343ba8c54800ac573d5e1caa2bfa200" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.155335 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.178271 4809 generic.go:334] "Generic (PLEG): container finished" podID="f91f6ab4-0c7c-4002-aad9-dcd408c9000e" containerID="d6cfe70d34456cc68d198ac9f1cb919fdebac98b1adbc19e4e8accb2affa5c81" exitCode=0 Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.178312 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f91f6ab4-0c7c-4002-aad9-dcd408c9000e","Type":"ContainerDied","Data":"d6cfe70d34456cc68d198ac9f1cb919fdebac98b1adbc19e4e8accb2affa5c81"} Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.212084 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.213955 4809 scope.go:117] "RemoveContainer" containerID="3e28454c296359bd8a170cd1fd9c49436203c71b0aad0ea763826736402a4456" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.237452 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.246230 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:15:22 crc kubenswrapper[4809]: E1124 07:15:22.246712 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d0a537c-b1f0-4908-a604-424ea6a58438" containerName="setup-container" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.246726 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d0a537c-b1f0-4908-a604-424ea6a58438" containerName="setup-container" Nov 24 07:15:22 crc kubenswrapper[4809]: E1124 07:15:22.246751 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d0a537c-b1f0-4908-a604-424ea6a58438" containerName="rabbitmq" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.246757 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d0a537c-b1f0-4908-a604-424ea6a58438" containerName="rabbitmq" Nov 24 07:15:22 crc kubenswrapper[4809]: E1124 07:15:22.246771 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b19ae12c-c3c5-4082-be4f-ffde1d1f56f0" containerName="collect-profiles" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.246899 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b19ae12c-c3c5-4082-be4f-ffde1d1f56f0" containerName="collect-profiles" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.247113 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b19ae12c-c3c5-4082-be4f-ffde1d1f56f0" containerName="collect-profiles" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.247142 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d0a537c-b1f0-4908-a604-424ea6a58438" containerName="rabbitmq" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.248823 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.252450 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-xfhq2" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.252642 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.252788 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.252882 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.253032 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.253132 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.259058 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.286443 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.391479 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.391563 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-config-data\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.391586 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.391618 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.391639 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krqcw\" (UniqueName: \"kubernetes.io/projected/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-kube-api-access-krqcw\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.391686 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.391725 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.391740 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.391759 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.391772 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.391792 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.493630 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.494602 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-config-data\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.495168 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-config-data\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.495251 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.495320 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.495358 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krqcw\" (UniqueName: \"kubernetes.io/projected/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-kube-api-access-krqcw\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.495501 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.495620 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.495642 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.495673 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.495690 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.495727 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.495754 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.496284 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.496503 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.496822 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.497134 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.497811 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.501987 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.504987 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.505080 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.505519 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.529334 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.544050 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krqcw\" (UniqueName: \"kubernetes.io/projected/4a1ec853-3d6a-403a-b2fe-82dc89ac5779-kube-api-access-krqcw\") pod \"rabbitmq-server-0\" (UID: \"4a1ec853-3d6a-403a-b2fe-82dc89ac5779\") " pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.597399 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-pod-info\") pod \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.597536 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-confd\") pod \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.597574 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-tls\") pod \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.597687 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-server-conf\") pod \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.597718 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-plugins\") pod \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.597852 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnlc6\" (UniqueName: \"kubernetes.io/projected/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-kube-api-access-lnlc6\") pod \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.597888 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-erlang-cookie\") pod \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.597974 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-erlang-cookie-secret\") pod \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.598036 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-plugins-conf\") pod \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.598089 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.598115 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-config-data\") pod \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\" (UID: \"f91f6ab4-0c7c-4002-aad9-dcd408c9000e\") " Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.598839 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.601084 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "f91f6ab4-0c7c-4002-aad9-dcd408c9000e" (UID: "f91f6ab4-0c7c-4002-aad9-dcd408c9000e"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.601282 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "f91f6ab4-0c7c-4002-aad9-dcd408c9000e" (UID: "f91f6ab4-0c7c-4002-aad9-dcd408c9000e"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.602163 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "f91f6ab4-0c7c-4002-aad9-dcd408c9000e" (UID: "f91f6ab4-0c7c-4002-aad9-dcd408c9000e"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.602333 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-pod-info" (OuterVolumeSpecName: "pod-info") pod "f91f6ab4-0c7c-4002-aad9-dcd408c9000e" (UID: "f91f6ab4-0c7c-4002-aad9-dcd408c9000e"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.606049 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-kube-api-access-lnlc6" (OuterVolumeSpecName: "kube-api-access-lnlc6") pod "f91f6ab4-0c7c-4002-aad9-dcd408c9000e" (UID: "f91f6ab4-0c7c-4002-aad9-dcd408c9000e"). InnerVolumeSpecName "kube-api-access-lnlc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.606751 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "f91f6ab4-0c7c-4002-aad9-dcd408c9000e" (UID: "f91f6ab4-0c7c-4002-aad9-dcd408c9000e"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.609503 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "f91f6ab4-0c7c-4002-aad9-dcd408c9000e" (UID: "f91f6ab4-0c7c-4002-aad9-dcd408c9000e"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.612128 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "f91f6ab4-0c7c-4002-aad9-dcd408c9000e" (UID: "f91f6ab4-0c7c-4002-aad9-dcd408c9000e"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.630687 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-config-data" (OuterVolumeSpecName: "config-data") pod "f91f6ab4-0c7c-4002-aad9-dcd408c9000e" (UID: "f91f6ab4-0c7c-4002-aad9-dcd408c9000e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.656313 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-server-conf" (OuterVolumeSpecName: "server-conf") pod "f91f6ab4-0c7c-4002-aad9-dcd408c9000e" (UID: "f91f6ab4-0c7c-4002-aad9-dcd408c9000e"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.700945 4809 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.701009 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.701022 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnlc6\" (UniqueName: \"kubernetes.io/projected/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-kube-api-access-lnlc6\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.701032 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.701041 4809 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.701050 4809 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.701246 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.701418 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.701612 4809 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.701888 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.716137 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "f91f6ab4-0c7c-4002-aad9-dcd408c9000e" (UID: "f91f6ab4-0c7c-4002-aad9-dcd408c9000e"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.722815 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.803778 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f91f6ab4-0c7c-4002-aad9-dcd408c9000e-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.804954 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:22 crc kubenswrapper[4809]: I1124 07:15:22.901680 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d0a537c-b1f0-4908-a604-424ea6a58438" path="/var/lib/kubelet/pods/7d0a537c-b1f0-4908-a604-424ea6a58438/volumes" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.038955 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.188219 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4a1ec853-3d6a-403a-b2fe-82dc89ac5779","Type":"ContainerStarted","Data":"ffb184c311a1a32f4ed6c43a378e6ac8d91c7455cc10eacd91f2c78c4bd83f24"} Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.191503 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f91f6ab4-0c7c-4002-aad9-dcd408c9000e","Type":"ContainerDied","Data":"d661a253ac318aced47d491b5aefdb4281e8dec6c3f04fff943e9ad0ef6cbf97"} Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.191555 4809 scope.go:117] "RemoveContainer" containerID="d6cfe70d34456cc68d198ac9f1cb919fdebac98b1adbc19e4e8accb2affa5c81" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.191651 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.225670 4809 scope.go:117] "RemoveContainer" containerID="ae51b70d41f1cc092e89e2f506b0d859bd045728a8e37fb6553593c1785b0da8" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.230199 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.245752 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.256319 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:15:23 crc kubenswrapper[4809]: E1124 07:15:23.256696 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f91f6ab4-0c7c-4002-aad9-dcd408c9000e" containerName="rabbitmq" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.256709 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f91f6ab4-0c7c-4002-aad9-dcd408c9000e" containerName="rabbitmq" Nov 24 07:15:23 crc kubenswrapper[4809]: E1124 07:15:23.256735 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f91f6ab4-0c7c-4002-aad9-dcd408c9000e" containerName="setup-container" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.256741 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f91f6ab4-0c7c-4002-aad9-dcd408c9000e" containerName="setup-container" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.256944 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f91f6ab4-0c7c-4002-aad9-dcd408c9000e" containerName="rabbitmq" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.258058 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.260648 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.260752 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.260856 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.260908 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.260998 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-ncq9f" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.261077 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.261377 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.273951 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.418702 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3285fa33-509a-4959-8547-0332857328c9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.419052 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3285fa33-509a-4959-8547-0332857328c9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.419137 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3285fa33-509a-4959-8547-0332857328c9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.419218 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfd5l\" (UniqueName: \"kubernetes.io/projected/3285fa33-509a-4959-8547-0332857328c9-kube-api-access-bfd5l\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.419356 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3285fa33-509a-4959-8547-0332857328c9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.419518 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.419580 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3285fa33-509a-4959-8547-0332857328c9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.419653 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3285fa33-509a-4959-8547-0332857328c9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.419764 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3285fa33-509a-4959-8547-0332857328c9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.419894 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3285fa33-509a-4959-8547-0332857328c9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.419928 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3285fa33-509a-4959-8547-0332857328c9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.521430 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3285fa33-509a-4959-8547-0332857328c9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.521522 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3285fa33-509a-4959-8547-0332857328c9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.521543 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3285fa33-509a-4959-8547-0332857328c9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.521563 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfd5l\" (UniqueName: \"kubernetes.io/projected/3285fa33-509a-4959-8547-0332857328c9-kube-api-access-bfd5l\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.521582 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3285fa33-509a-4959-8547-0332857328c9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.521628 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.521653 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3285fa33-509a-4959-8547-0332857328c9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.521684 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3285fa33-509a-4959-8547-0332857328c9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.521711 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3285fa33-509a-4959-8547-0332857328c9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.521752 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3285fa33-509a-4959-8547-0332857328c9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.521770 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3285fa33-509a-4959-8547-0332857328c9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.521911 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3285fa33-509a-4959-8547-0332857328c9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.522262 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.522946 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3285fa33-509a-4959-8547-0332857328c9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.523219 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3285fa33-509a-4959-8547-0332857328c9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.523802 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3285fa33-509a-4959-8547-0332857328c9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.523889 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3285fa33-509a-4959-8547-0332857328c9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.526997 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3285fa33-509a-4959-8547-0332857328c9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.527188 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3285fa33-509a-4959-8547-0332857328c9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.527343 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3285fa33-509a-4959-8547-0332857328c9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.531463 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3285fa33-509a-4959-8547-0332857328c9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.540209 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfd5l\" (UniqueName: \"kubernetes.io/projected/3285fa33-509a-4959-8547-0332857328c9-kube-api-access-bfd5l\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.556806 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3285fa33-509a-4959-8547-0332857328c9\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.623551 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.780578 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-tz7zj"] Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.782776 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.786480 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.795504 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-tz7zj"] Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.931446 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pzp8\" (UniqueName: \"kubernetes.io/projected/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-kube-api-access-4pzp8\") pod \"dnsmasq-dns-79bd4cc8c9-tz7zj\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.931490 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-tz7zj\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.931535 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-tz7zj\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.931592 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-config\") pod \"dnsmasq-dns-79bd4cc8c9-tz7zj\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.931619 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-tz7zj\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.931690 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-tz7zj\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:23 crc kubenswrapper[4809]: I1124 07:15:23.931756 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-tz7zj\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:24 crc kubenswrapper[4809]: I1124 07:15:24.033646 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-config\") pod \"dnsmasq-dns-79bd4cc8c9-tz7zj\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:24 crc kubenswrapper[4809]: I1124 07:15:24.033696 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-tz7zj\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:24 crc kubenswrapper[4809]: I1124 07:15:24.033779 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-tz7zj\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:24 crc kubenswrapper[4809]: I1124 07:15:24.033801 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-tz7zj\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:24 crc kubenswrapper[4809]: I1124 07:15:24.033851 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pzp8\" (UniqueName: \"kubernetes.io/projected/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-kube-api-access-4pzp8\") pod \"dnsmasq-dns-79bd4cc8c9-tz7zj\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:24 crc kubenswrapper[4809]: I1124 07:15:24.033873 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-tz7zj\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:24 crc kubenswrapper[4809]: I1124 07:15:24.033908 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-tz7zj\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:24 crc kubenswrapper[4809]: I1124 07:15:24.034745 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-tz7zj\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:24 crc kubenswrapper[4809]: I1124 07:15:24.035395 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-config\") pod \"dnsmasq-dns-79bd4cc8c9-tz7zj\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:24 crc kubenswrapper[4809]: I1124 07:15:24.036463 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-tz7zj\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:24 crc kubenswrapper[4809]: I1124 07:15:24.036578 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-tz7zj\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:24 crc kubenswrapper[4809]: I1124 07:15:24.037017 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-tz7zj\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:24 crc kubenswrapper[4809]: I1124 07:15:24.037025 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-tz7zj\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:24 crc kubenswrapper[4809]: I1124 07:15:24.058165 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pzp8\" (UniqueName: \"kubernetes.io/projected/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-kube-api-access-4pzp8\") pod \"dnsmasq-dns-79bd4cc8c9-tz7zj\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:24 crc kubenswrapper[4809]: I1124 07:15:24.105695 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:24 crc kubenswrapper[4809]: I1124 07:15:24.120282 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:15:24 crc kubenswrapper[4809]: W1124 07:15:24.123644 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3285fa33_509a_4959_8547_0332857328c9.slice/crio-5f5970177b813134fecfc3eaa7f7ee1347d9b803deb31c0c8b6db5fe10fe6671 WatchSource:0}: Error finding container 5f5970177b813134fecfc3eaa7f7ee1347d9b803deb31c0c8b6db5fe10fe6671: Status 404 returned error can't find the container with id 5f5970177b813134fecfc3eaa7f7ee1347d9b803deb31c0c8b6db5fe10fe6671 Nov 24 07:15:24 crc kubenswrapper[4809]: I1124 07:15:24.244610 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3285fa33-509a-4959-8547-0332857328c9","Type":"ContainerStarted","Data":"5f5970177b813134fecfc3eaa7f7ee1347d9b803deb31c0c8b6db5fe10fe6671"} Nov 24 07:15:24 crc kubenswrapper[4809]: W1124 07:15:24.583917 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2b43e96_a08d_4335_bbe5_3aded32e3b7a.slice/crio-db2e9025246b336958435ae6c065b3ef1f5833e865fee95af83b0cc806437d4e WatchSource:0}: Error finding container db2e9025246b336958435ae6c065b3ef1f5833e865fee95af83b0cc806437d4e: Status 404 returned error can't find the container with id db2e9025246b336958435ae6c065b3ef1f5833e865fee95af83b0cc806437d4e Nov 24 07:15:24 crc kubenswrapper[4809]: I1124 07:15:24.593378 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-tz7zj"] Nov 24 07:15:24 crc kubenswrapper[4809]: I1124 07:15:24.904895 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f91f6ab4-0c7c-4002-aad9-dcd408c9000e" path="/var/lib/kubelet/pods/f91f6ab4-0c7c-4002-aad9-dcd408c9000e/volumes" Nov 24 07:15:25 crc kubenswrapper[4809]: I1124 07:15:25.254434 4809 generic.go:334] "Generic (PLEG): container finished" podID="c2b43e96-a08d-4335-bbe5-3aded32e3b7a" containerID="bd0327e4af35df63870e1f69945fd1979dcf0ce2ec0083a5c178d41eedada22c" exitCode=0 Nov 24 07:15:25 crc kubenswrapper[4809]: I1124 07:15:25.254561 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" event={"ID":"c2b43e96-a08d-4335-bbe5-3aded32e3b7a","Type":"ContainerDied","Data":"bd0327e4af35df63870e1f69945fd1979dcf0ce2ec0083a5c178d41eedada22c"} Nov 24 07:15:25 crc kubenswrapper[4809]: I1124 07:15:25.254909 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" event={"ID":"c2b43e96-a08d-4335-bbe5-3aded32e3b7a","Type":"ContainerStarted","Data":"db2e9025246b336958435ae6c065b3ef1f5833e865fee95af83b0cc806437d4e"} Nov 24 07:15:25 crc kubenswrapper[4809]: I1124 07:15:25.258527 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4a1ec853-3d6a-403a-b2fe-82dc89ac5779","Type":"ContainerStarted","Data":"6bf7e4bc197239141c07ac0b34fbc2a8668f2a38e66cb0514df0bf0190b1f3f0"} Nov 24 07:15:26 crc kubenswrapper[4809]: I1124 07:15:26.270939 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3285fa33-509a-4959-8547-0332857328c9","Type":"ContainerStarted","Data":"3d5711f4ea492904e688de95c7dd66a53ce5c02b15ed3aa1ca5bd3da4e438bef"} Nov 24 07:15:26 crc kubenswrapper[4809]: I1124 07:15:26.275195 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" event={"ID":"c2b43e96-a08d-4335-bbe5-3aded32e3b7a","Type":"ContainerStarted","Data":"3410d40cb3ead530cdf54702dabd4c8c3358e7521efa78b750654f53075cfe60"} Nov 24 07:15:26 crc kubenswrapper[4809]: I1124 07:15:26.275274 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:26 crc kubenswrapper[4809]: I1124 07:15:26.323290 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" podStartSLOduration=3.32327114 podStartE2EDuration="3.32327114s" podCreationTimestamp="2025-11-24 07:15:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:15:26.317216339 +0000 UTC m=+1226.217808064" watchObservedRunningTime="2025-11-24 07:15:26.32327114 +0000 UTC m=+1226.223862855" Nov 24 07:15:34 crc kubenswrapper[4809]: I1124 07:15:34.107818 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:34 crc kubenswrapper[4809]: I1124 07:15:34.181425 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-w774p"] Nov 24 07:15:34 crc kubenswrapper[4809]: I1124 07:15:34.181935 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" podUID="e535ca4c-f961-44f0-a031-61d4264a984c" containerName="dnsmasq-dns" containerID="cri-o://f9ff0326dfa3960ea21b7f760328dcba59a13ea8bfdf0dcdd7c217a9570755e0" gracePeriod=10 Nov 24 07:15:34 crc kubenswrapper[4809]: I1124 07:15:34.352664 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55478c4467-f6c2x"] Nov 24 07:15:34 crc kubenswrapper[4809]: I1124 07:15:34.354400 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:34 crc kubenswrapper[4809]: I1124 07:15:34.366435 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-f6c2x"] Nov 24 07:15:34 crc kubenswrapper[4809]: I1124 07:15:34.396782 4809 generic.go:334] "Generic (PLEG): container finished" podID="e535ca4c-f961-44f0-a031-61d4264a984c" containerID="f9ff0326dfa3960ea21b7f760328dcba59a13ea8bfdf0dcdd7c217a9570755e0" exitCode=0 Nov 24 07:15:34 crc kubenswrapper[4809]: I1124 07:15:34.396835 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" event={"ID":"e535ca4c-f961-44f0-a031-61d4264a984c","Type":"ContainerDied","Data":"f9ff0326dfa3960ea21b7f760328dcba59a13ea8bfdf0dcdd7c217a9570755e0"} Nov 24 07:15:34 crc kubenswrapper[4809]: I1124 07:15:34.445214 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09eb3227-9b7b-4d9a-9fa1-80e28696c61e-dns-svc\") pod \"dnsmasq-dns-55478c4467-f6c2x\" (UID: \"09eb3227-9b7b-4d9a-9fa1-80e28696c61e\") " pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:34 crc kubenswrapper[4809]: I1124 07:15:34.445272 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/09eb3227-9b7b-4d9a-9fa1-80e28696c61e-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-f6c2x\" (UID: \"09eb3227-9b7b-4d9a-9fa1-80e28696c61e\") " pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:34 crc kubenswrapper[4809]: I1124 07:15:34.445309 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09eb3227-9b7b-4d9a-9fa1-80e28696c61e-config\") pod \"dnsmasq-dns-55478c4467-f6c2x\" (UID: \"09eb3227-9b7b-4d9a-9fa1-80e28696c61e\") " pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:34 crc kubenswrapper[4809]: I1124 07:15:34.445334 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/09eb3227-9b7b-4d9a-9fa1-80e28696c61e-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-f6c2x\" (UID: \"09eb3227-9b7b-4d9a-9fa1-80e28696c61e\") " pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:34 crc kubenswrapper[4809]: I1124 07:15:34.445379 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/09eb3227-9b7b-4d9a-9fa1-80e28696c61e-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-f6c2x\" (UID: \"09eb3227-9b7b-4d9a-9fa1-80e28696c61e\") " pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:34 crc kubenswrapper[4809]: I1124 07:15:34.445434 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09eb3227-9b7b-4d9a-9fa1-80e28696c61e-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-f6c2x\" (UID: \"09eb3227-9b7b-4d9a-9fa1-80e28696c61e\") " pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:34 crc kubenswrapper[4809]: I1124 07:15:34.445586 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76wtz\" (UniqueName: \"kubernetes.io/projected/09eb3227-9b7b-4d9a-9fa1-80e28696c61e-kube-api-access-76wtz\") pod \"dnsmasq-dns-55478c4467-f6c2x\" (UID: \"09eb3227-9b7b-4d9a-9fa1-80e28696c61e\") " pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.546956 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09eb3227-9b7b-4d9a-9fa1-80e28696c61e-dns-svc\") pod \"dnsmasq-dns-55478c4467-f6c2x\" (UID: \"09eb3227-9b7b-4d9a-9fa1-80e28696c61e\") " pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.547316 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/09eb3227-9b7b-4d9a-9fa1-80e28696c61e-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-f6c2x\" (UID: \"09eb3227-9b7b-4d9a-9fa1-80e28696c61e\") " pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.547347 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09eb3227-9b7b-4d9a-9fa1-80e28696c61e-config\") pod \"dnsmasq-dns-55478c4467-f6c2x\" (UID: \"09eb3227-9b7b-4d9a-9fa1-80e28696c61e\") " pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.547370 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/09eb3227-9b7b-4d9a-9fa1-80e28696c61e-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-f6c2x\" (UID: \"09eb3227-9b7b-4d9a-9fa1-80e28696c61e\") " pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.548166 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09eb3227-9b7b-4d9a-9fa1-80e28696c61e-config\") pod \"dnsmasq-dns-55478c4467-f6c2x\" (UID: \"09eb3227-9b7b-4d9a-9fa1-80e28696c61e\") " pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.548200 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/09eb3227-9b7b-4d9a-9fa1-80e28696c61e-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-f6c2x\" (UID: \"09eb3227-9b7b-4d9a-9fa1-80e28696c61e\") " pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.548286 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09eb3227-9b7b-4d9a-9fa1-80e28696c61e-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-f6c2x\" (UID: \"09eb3227-9b7b-4d9a-9fa1-80e28696c61e\") " pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.548515 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/09eb3227-9b7b-4d9a-9fa1-80e28696c61e-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-f6c2x\" (UID: \"09eb3227-9b7b-4d9a-9fa1-80e28696c61e\") " pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.548594 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/09eb3227-9b7b-4d9a-9fa1-80e28696c61e-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-f6c2x\" (UID: \"09eb3227-9b7b-4d9a-9fa1-80e28696c61e\") " pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.548838 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/09eb3227-9b7b-4d9a-9fa1-80e28696c61e-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-f6c2x\" (UID: \"09eb3227-9b7b-4d9a-9fa1-80e28696c61e\") " pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.548886 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09eb3227-9b7b-4d9a-9fa1-80e28696c61e-dns-svc\") pod \"dnsmasq-dns-55478c4467-f6c2x\" (UID: \"09eb3227-9b7b-4d9a-9fa1-80e28696c61e\") " pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.549343 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09eb3227-9b7b-4d9a-9fa1-80e28696c61e-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-f6c2x\" (UID: \"09eb3227-9b7b-4d9a-9fa1-80e28696c61e\") " pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.549520 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76wtz\" (UniqueName: \"kubernetes.io/projected/09eb3227-9b7b-4d9a-9fa1-80e28696c61e-kube-api-access-76wtz\") pod \"dnsmasq-dns-55478c4467-f6c2x\" (UID: \"09eb3227-9b7b-4d9a-9fa1-80e28696c61e\") " pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.588114 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76wtz\" (UniqueName: \"kubernetes.io/projected/09eb3227-9b7b-4d9a-9fa1-80e28696c61e-kube-api-access-76wtz\") pod \"dnsmasq-dns-55478c4467-f6c2x\" (UID: \"09eb3227-9b7b-4d9a-9fa1-80e28696c61e\") " pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.692642 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.774073 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.855606 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-ovsdbserver-sb\") pod \"e535ca4c-f961-44f0-a031-61d4264a984c\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.855742 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-ovsdbserver-nb\") pod \"e535ca4c-f961-44f0-a031-61d4264a984c\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.855814 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-dns-svc\") pod \"e535ca4c-f961-44f0-a031-61d4264a984c\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.855937 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-config\") pod \"e535ca4c-f961-44f0-a031-61d4264a984c\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.856004 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-dns-swift-storage-0\") pod \"e535ca4c-f961-44f0-a031-61d4264a984c\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.856050 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwmf4\" (UniqueName: \"kubernetes.io/projected/e535ca4c-f961-44f0-a031-61d4264a984c-kube-api-access-cwmf4\") pod \"e535ca4c-f961-44f0-a031-61d4264a984c\" (UID: \"e535ca4c-f961-44f0-a031-61d4264a984c\") " Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.860152 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e535ca4c-f961-44f0-a031-61d4264a984c-kube-api-access-cwmf4" (OuterVolumeSpecName: "kube-api-access-cwmf4") pod "e535ca4c-f961-44f0-a031-61d4264a984c" (UID: "e535ca4c-f961-44f0-a031-61d4264a984c"). InnerVolumeSpecName "kube-api-access-cwmf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.905022 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e535ca4c-f961-44f0-a031-61d4264a984c" (UID: "e535ca4c-f961-44f0-a031-61d4264a984c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.929571 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e535ca4c-f961-44f0-a031-61d4264a984c" (UID: "e535ca4c-f961-44f0-a031-61d4264a984c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.959190 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.959237 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwmf4\" (UniqueName: \"kubernetes.io/projected/e535ca4c-f961-44f0-a031-61d4264a984c-kube-api-access-cwmf4\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.959251 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.967908 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-config" (OuterVolumeSpecName: "config") pod "e535ca4c-f961-44f0-a031-61d4264a984c" (UID: "e535ca4c-f961-44f0-a031-61d4264a984c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:34.968727 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e535ca4c-f961-44f0-a031-61d4264a984c" (UID: "e535ca4c-f961-44f0-a031-61d4264a984c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:35.000784 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e535ca4c-f961-44f0-a031-61d4264a984c" (UID: "e535ca4c-f961-44f0-a031-61d4264a984c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:35.061261 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:35.061292 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:35.061307 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e535ca4c-f961-44f0-a031-61d4264a984c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:35.412146 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" event={"ID":"e535ca4c-f961-44f0-a031-61d4264a984c","Type":"ContainerDied","Data":"66c03ee2d9f6909b36e505a5fc232e3106e1493ea382fa35ee00f5f33e959163"} Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:35.412397 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-w774p" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:35.412504 4809 scope.go:117] "RemoveContainer" containerID="f9ff0326dfa3960ea21b7f760328dcba59a13ea8bfdf0dcdd7c217a9570755e0" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:35.452527 4809 scope.go:117] "RemoveContainer" containerID="4b795101827cd4186f9dda1abb76897656daeedd6ece3f3f35b3eb0f465b0638" Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:35.460527 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-w774p"] Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:35.470616 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-w774p"] Nov 24 07:15:35 crc kubenswrapper[4809]: I1124 07:15:35.629306 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-f6c2x"] Nov 24 07:15:36 crc kubenswrapper[4809]: I1124 07:15:36.423820 4809 generic.go:334] "Generic (PLEG): container finished" podID="09eb3227-9b7b-4d9a-9fa1-80e28696c61e" containerID="f5f38b5d9056f4d7f6af973eb07f5902304d2d5de57617bdc5bd1bdc0a5065b4" exitCode=0 Nov 24 07:15:36 crc kubenswrapper[4809]: I1124 07:15:36.423881 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-f6c2x" event={"ID":"09eb3227-9b7b-4d9a-9fa1-80e28696c61e","Type":"ContainerDied","Data":"f5f38b5d9056f4d7f6af973eb07f5902304d2d5de57617bdc5bd1bdc0a5065b4"} Nov 24 07:15:36 crc kubenswrapper[4809]: I1124 07:15:36.424222 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-f6c2x" event={"ID":"09eb3227-9b7b-4d9a-9fa1-80e28696c61e","Type":"ContainerStarted","Data":"5b690a82cc7897225b12a0d2901cd509c33e493181393fc94fe66b5501f2454d"} Nov 24 07:15:36 crc kubenswrapper[4809]: I1124 07:15:36.904170 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e535ca4c-f961-44f0-a031-61d4264a984c" path="/var/lib/kubelet/pods/e535ca4c-f961-44f0-a031-61d4264a984c/volumes" Nov 24 07:15:37 crc kubenswrapper[4809]: I1124 07:15:37.440637 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-f6c2x" event={"ID":"09eb3227-9b7b-4d9a-9fa1-80e28696c61e","Type":"ContainerStarted","Data":"15765ddde2ed26e1d95081db9408b7a53bf13e910d6a675417f1b550fb290c07"} Nov 24 07:15:37 crc kubenswrapper[4809]: I1124 07:15:37.440821 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:37 crc kubenswrapper[4809]: I1124 07:15:37.468594 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55478c4467-f6c2x" podStartSLOduration=3.468571322 podStartE2EDuration="3.468571322s" podCreationTimestamp="2025-11-24 07:15:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:15:37.460939538 +0000 UTC m=+1237.361531253" watchObservedRunningTime="2025-11-24 07:15:37.468571322 +0000 UTC m=+1237.369163037" Nov 24 07:15:44 crc kubenswrapper[4809]: I1124 07:15:44.694187 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55478c4467-f6c2x" Nov 24 07:15:44 crc kubenswrapper[4809]: I1124 07:15:44.767960 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-tz7zj"] Nov 24 07:15:44 crc kubenswrapper[4809]: I1124 07:15:44.768590 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" podUID="c2b43e96-a08d-4335-bbe5-3aded32e3b7a" containerName="dnsmasq-dns" containerID="cri-o://3410d40cb3ead530cdf54702dabd4c8c3358e7521efa78b750654f53075cfe60" gracePeriod=10 Nov 24 07:15:45 crc kubenswrapper[4809]: I1124 07:15:45.523678 4809 generic.go:334] "Generic (PLEG): container finished" podID="c2b43e96-a08d-4335-bbe5-3aded32e3b7a" containerID="3410d40cb3ead530cdf54702dabd4c8c3358e7521efa78b750654f53075cfe60" exitCode=0 Nov 24 07:15:45 crc kubenswrapper[4809]: I1124 07:15:45.524035 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" event={"ID":"c2b43e96-a08d-4335-bbe5-3aded32e3b7a","Type":"ContainerDied","Data":"3410d40cb3ead530cdf54702dabd4c8c3358e7521efa78b750654f53075cfe60"} Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.361750 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.487231 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-ovsdbserver-sb\") pod \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.487312 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-ovsdbserver-nb\") pod \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.487474 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pzp8\" (UniqueName: \"kubernetes.io/projected/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-kube-api-access-4pzp8\") pod \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.487547 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-openstack-edpm-ipam\") pod \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.487612 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-config\") pod \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.487665 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-dns-swift-storage-0\") pod \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.487898 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-dns-svc\") pod \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\" (UID: \"c2b43e96-a08d-4335-bbe5-3aded32e3b7a\") " Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.494454 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-kube-api-access-4pzp8" (OuterVolumeSpecName: "kube-api-access-4pzp8") pod "c2b43e96-a08d-4335-bbe5-3aded32e3b7a" (UID: "c2b43e96-a08d-4335-bbe5-3aded32e3b7a"). InnerVolumeSpecName "kube-api-access-4pzp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.536669 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" event={"ID":"c2b43e96-a08d-4335-bbe5-3aded32e3b7a","Type":"ContainerDied","Data":"db2e9025246b336958435ae6c065b3ef1f5833e865fee95af83b0cc806437d4e"} Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.536722 4809 scope.go:117] "RemoveContainer" containerID="3410d40cb3ead530cdf54702dabd4c8c3358e7521efa78b750654f53075cfe60" Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.536772 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-tz7zj" Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.544225 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c2b43e96-a08d-4335-bbe5-3aded32e3b7a" (UID: "c2b43e96-a08d-4335-bbe5-3aded32e3b7a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.548932 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "c2b43e96-a08d-4335-bbe5-3aded32e3b7a" (UID: "c2b43e96-a08d-4335-bbe5-3aded32e3b7a"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.551522 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c2b43e96-a08d-4335-bbe5-3aded32e3b7a" (UID: "c2b43e96-a08d-4335-bbe5-3aded32e3b7a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.554228 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c2b43e96-a08d-4335-bbe5-3aded32e3b7a" (UID: "c2b43e96-a08d-4335-bbe5-3aded32e3b7a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.560337 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c2b43e96-a08d-4335-bbe5-3aded32e3b7a" (UID: "c2b43e96-a08d-4335-bbe5-3aded32e3b7a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.562167 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-config" (OuterVolumeSpecName: "config") pod "c2b43e96-a08d-4335-bbe5-3aded32e3b7a" (UID: "c2b43e96-a08d-4335-bbe5-3aded32e3b7a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.590918 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.590960 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.590995 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.591008 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pzp8\" (UniqueName: \"kubernetes.io/projected/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-kube-api-access-4pzp8\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.591020 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.591032 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.591043 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c2b43e96-a08d-4335-bbe5-3aded32e3b7a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.664624 4809 scope.go:117] "RemoveContainer" containerID="bd0327e4af35df63870e1f69945fd1979dcf0ce2ec0083a5c178d41eedada22c" Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.876612 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-tz7zj"] Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.883997 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-tz7zj"] Nov 24 07:15:46 crc kubenswrapper[4809]: I1124 07:15:46.904489 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2b43e96-a08d-4335-bbe5-3aded32e3b7a" path="/var/lib/kubelet/pods/c2b43e96-a08d-4335-bbe5-3aded32e3b7a/volumes" Nov 24 07:15:48 crc kubenswrapper[4809]: I1124 07:15:48.043736 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:15:48 crc kubenswrapper[4809]: I1124 07:15:48.044153 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.403458 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9"] Nov 24 07:15:53 crc kubenswrapper[4809]: E1124 07:15:53.404333 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e535ca4c-f961-44f0-a031-61d4264a984c" containerName="dnsmasq-dns" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.404347 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e535ca4c-f961-44f0-a031-61d4264a984c" containerName="dnsmasq-dns" Nov 24 07:15:53 crc kubenswrapper[4809]: E1124 07:15:53.404368 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e535ca4c-f961-44f0-a031-61d4264a984c" containerName="init" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.404376 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e535ca4c-f961-44f0-a031-61d4264a984c" containerName="init" Nov 24 07:15:53 crc kubenswrapper[4809]: E1124 07:15:53.404388 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2b43e96-a08d-4335-bbe5-3aded32e3b7a" containerName="dnsmasq-dns" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.404395 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2b43e96-a08d-4335-bbe5-3aded32e3b7a" containerName="dnsmasq-dns" Nov 24 07:15:53 crc kubenswrapper[4809]: E1124 07:15:53.404437 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2b43e96-a08d-4335-bbe5-3aded32e3b7a" containerName="init" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.404445 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2b43e96-a08d-4335-bbe5-3aded32e3b7a" containerName="init" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.404623 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2b43e96-a08d-4335-bbe5-3aded32e3b7a" containerName="dnsmasq-dns" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.404645 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e535ca4c-f961-44f0-a031-61d4264a984c" containerName="dnsmasq-dns" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.405285 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.407415 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.407748 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.407935 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lx7ll" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.415589 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.417938 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9"] Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.423897 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7ql9\" (UniqueName: \"kubernetes.io/projected/9ed9735b-66b5-44b3-9399-af938bfcd802-kube-api-access-n7ql9\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9\" (UID: \"9ed9735b-66b5-44b3-9399-af938bfcd802\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.424239 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ed9735b-66b5-44b3-9399-af938bfcd802-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9\" (UID: \"9ed9735b-66b5-44b3-9399-af938bfcd802\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.424355 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ed9735b-66b5-44b3-9399-af938bfcd802-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9\" (UID: \"9ed9735b-66b5-44b3-9399-af938bfcd802\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.424472 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ed9735b-66b5-44b3-9399-af938bfcd802-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9\" (UID: \"9ed9735b-66b5-44b3-9399-af938bfcd802\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.526308 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7ql9\" (UniqueName: \"kubernetes.io/projected/9ed9735b-66b5-44b3-9399-af938bfcd802-kube-api-access-n7ql9\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9\" (UID: \"9ed9735b-66b5-44b3-9399-af938bfcd802\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.526454 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ed9735b-66b5-44b3-9399-af938bfcd802-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9\" (UID: \"9ed9735b-66b5-44b3-9399-af938bfcd802\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.526478 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ed9735b-66b5-44b3-9399-af938bfcd802-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9\" (UID: \"9ed9735b-66b5-44b3-9399-af938bfcd802\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.526499 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ed9735b-66b5-44b3-9399-af938bfcd802-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9\" (UID: \"9ed9735b-66b5-44b3-9399-af938bfcd802\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.539774 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ed9735b-66b5-44b3-9399-af938bfcd802-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9\" (UID: \"9ed9735b-66b5-44b3-9399-af938bfcd802\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.539780 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ed9735b-66b5-44b3-9399-af938bfcd802-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9\" (UID: \"9ed9735b-66b5-44b3-9399-af938bfcd802\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.540220 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ed9735b-66b5-44b3-9399-af938bfcd802-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9\" (UID: \"9ed9735b-66b5-44b3-9399-af938bfcd802\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.561107 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7ql9\" (UniqueName: \"kubernetes.io/projected/9ed9735b-66b5-44b3-9399-af938bfcd802-kube-api-access-n7ql9\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9\" (UID: \"9ed9735b-66b5-44b3-9399-af938bfcd802\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9" Nov 24 07:15:53 crc kubenswrapper[4809]: I1124 07:15:53.724769 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9" Nov 24 07:15:54 crc kubenswrapper[4809]: I1124 07:15:54.237164 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9"] Nov 24 07:15:54 crc kubenswrapper[4809]: I1124 07:15:54.248597 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 07:15:54 crc kubenswrapper[4809]: I1124 07:15:54.623954 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9" event={"ID":"9ed9735b-66b5-44b3-9399-af938bfcd802","Type":"ContainerStarted","Data":"a6b5ab4a189341699779a665e29a04a55c439b9fa458868f313b42000935b54d"} Nov 24 07:15:57 crc kubenswrapper[4809]: I1124 07:15:57.653806 4809 generic.go:334] "Generic (PLEG): container finished" podID="4a1ec853-3d6a-403a-b2fe-82dc89ac5779" containerID="6bf7e4bc197239141c07ac0b34fbc2a8668f2a38e66cb0514df0bf0190b1f3f0" exitCode=0 Nov 24 07:15:57 crc kubenswrapper[4809]: I1124 07:15:57.653918 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4a1ec853-3d6a-403a-b2fe-82dc89ac5779","Type":"ContainerDied","Data":"6bf7e4bc197239141c07ac0b34fbc2a8668f2a38e66cb0514df0bf0190b1f3f0"} Nov 24 07:15:58 crc kubenswrapper[4809]: I1124 07:15:58.670637 4809 generic.go:334] "Generic (PLEG): container finished" podID="3285fa33-509a-4959-8547-0332857328c9" containerID="3d5711f4ea492904e688de95c7dd66a53ce5c02b15ed3aa1ca5bd3da4e438bef" exitCode=0 Nov 24 07:15:58 crc kubenswrapper[4809]: I1124 07:15:58.670745 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3285fa33-509a-4959-8547-0332857328c9","Type":"ContainerDied","Data":"3d5711f4ea492904e688de95c7dd66a53ce5c02b15ed3aa1ca5bd3da4e438bef"} Nov 24 07:16:02 crc kubenswrapper[4809]: I1124 07:16:02.285891 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 07:16:02 crc kubenswrapper[4809]: I1124 07:16:02.713284 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3285fa33-509a-4959-8547-0332857328c9","Type":"ContainerStarted","Data":"0894c882e3cb87b43d59c2a93ec97e8c43c21de34e4899fdb2cce70cc4fa0b89"} Nov 24 07:16:02 crc kubenswrapper[4809]: I1124 07:16:02.714929 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:16:02 crc kubenswrapper[4809]: I1124 07:16:02.718835 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4a1ec853-3d6a-403a-b2fe-82dc89ac5779","Type":"ContainerStarted","Data":"58d1697b2d276f8d270b42e90bfe621bb433f56e56990b0f3fef8a691dc78708"} Nov 24 07:16:02 crc kubenswrapper[4809]: I1124 07:16:02.720217 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 07:16:02 crc kubenswrapper[4809]: I1124 07:16:02.722082 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9" event={"ID":"9ed9735b-66b5-44b3-9399-af938bfcd802","Type":"ContainerStarted","Data":"63f8701cca92c493a795153e52096f80c85244ded63b008777762905127e3132"} Nov 24 07:16:02 crc kubenswrapper[4809]: I1124 07:16:02.751677 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=39.751655965 podStartE2EDuration="39.751655965s" podCreationTimestamp="2025-11-24 07:15:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:16:02.745901671 +0000 UTC m=+1262.646493416" watchObservedRunningTime="2025-11-24 07:16:02.751655965 +0000 UTC m=+1262.652247680" Nov 24 07:16:02 crc kubenswrapper[4809]: I1124 07:16:02.787827 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=40.78780025 podStartE2EDuration="40.78780025s" podCreationTimestamp="2025-11-24 07:15:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:16:02.779148289 +0000 UTC m=+1262.679740024" watchObservedRunningTime="2025-11-24 07:16:02.78780025 +0000 UTC m=+1262.688391995" Nov 24 07:16:02 crc kubenswrapper[4809]: I1124 07:16:02.815996 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9" podStartSLOduration=1.7812840159999999 podStartE2EDuration="9.815950341s" podCreationTimestamp="2025-11-24 07:15:53 +0000 UTC" firstStartedPulling="2025-11-24 07:15:54.24837902 +0000 UTC m=+1254.148970725" lastFinishedPulling="2025-11-24 07:16:02.283045335 +0000 UTC m=+1262.183637050" observedRunningTime="2025-11-24 07:16:02.799608615 +0000 UTC m=+1262.700200320" watchObservedRunningTime="2025-11-24 07:16:02.815950341 +0000 UTC m=+1262.716542056" Nov 24 07:16:12 crc kubenswrapper[4809]: I1124 07:16:12.605217 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 07:16:13 crc kubenswrapper[4809]: I1124 07:16:13.627219 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:16:13 crc kubenswrapper[4809]: I1124 07:16:13.841377 4809 generic.go:334] "Generic (PLEG): container finished" podID="9ed9735b-66b5-44b3-9399-af938bfcd802" containerID="63f8701cca92c493a795153e52096f80c85244ded63b008777762905127e3132" exitCode=0 Nov 24 07:16:13 crc kubenswrapper[4809]: I1124 07:16:13.841431 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9" event={"ID":"9ed9735b-66b5-44b3-9399-af938bfcd802","Type":"ContainerDied","Data":"63f8701cca92c493a795153e52096f80c85244ded63b008777762905127e3132"} Nov 24 07:16:15 crc kubenswrapper[4809]: I1124 07:16:15.305540 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9" Nov 24 07:16:15 crc kubenswrapper[4809]: I1124 07:16:15.327659 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ed9735b-66b5-44b3-9399-af938bfcd802-inventory\") pod \"9ed9735b-66b5-44b3-9399-af938bfcd802\" (UID: \"9ed9735b-66b5-44b3-9399-af938bfcd802\") " Nov 24 07:16:15 crc kubenswrapper[4809]: I1124 07:16:15.327737 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ed9735b-66b5-44b3-9399-af938bfcd802-ssh-key\") pod \"9ed9735b-66b5-44b3-9399-af938bfcd802\" (UID: \"9ed9735b-66b5-44b3-9399-af938bfcd802\") " Nov 24 07:16:15 crc kubenswrapper[4809]: I1124 07:16:15.327845 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ed9735b-66b5-44b3-9399-af938bfcd802-repo-setup-combined-ca-bundle\") pod \"9ed9735b-66b5-44b3-9399-af938bfcd802\" (UID: \"9ed9735b-66b5-44b3-9399-af938bfcd802\") " Nov 24 07:16:15 crc kubenswrapper[4809]: I1124 07:16:15.327928 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7ql9\" (UniqueName: \"kubernetes.io/projected/9ed9735b-66b5-44b3-9399-af938bfcd802-kube-api-access-n7ql9\") pod \"9ed9735b-66b5-44b3-9399-af938bfcd802\" (UID: \"9ed9735b-66b5-44b3-9399-af938bfcd802\") " Nov 24 07:16:15 crc kubenswrapper[4809]: I1124 07:16:15.345418 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ed9735b-66b5-44b3-9399-af938bfcd802-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "9ed9735b-66b5-44b3-9399-af938bfcd802" (UID: "9ed9735b-66b5-44b3-9399-af938bfcd802"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:16:15 crc kubenswrapper[4809]: I1124 07:16:15.354049 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ed9735b-66b5-44b3-9399-af938bfcd802-kube-api-access-n7ql9" (OuterVolumeSpecName: "kube-api-access-n7ql9") pod "9ed9735b-66b5-44b3-9399-af938bfcd802" (UID: "9ed9735b-66b5-44b3-9399-af938bfcd802"). InnerVolumeSpecName "kube-api-access-n7ql9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:16:15 crc kubenswrapper[4809]: I1124 07:16:15.378547 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ed9735b-66b5-44b3-9399-af938bfcd802-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9ed9735b-66b5-44b3-9399-af938bfcd802" (UID: "9ed9735b-66b5-44b3-9399-af938bfcd802"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:16:15 crc kubenswrapper[4809]: I1124 07:16:15.430162 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ed9735b-66b5-44b3-9399-af938bfcd802-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 07:16:15 crc kubenswrapper[4809]: I1124 07:16:15.430192 4809 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ed9735b-66b5-44b3-9399-af938bfcd802-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:16:15 crc kubenswrapper[4809]: I1124 07:16:15.430208 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7ql9\" (UniqueName: \"kubernetes.io/projected/9ed9735b-66b5-44b3-9399-af938bfcd802-kube-api-access-n7ql9\") on node \"crc\" DevicePath \"\"" Nov 24 07:16:15 crc kubenswrapper[4809]: I1124 07:16:15.435091 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ed9735b-66b5-44b3-9399-af938bfcd802-inventory" (OuterVolumeSpecName: "inventory") pod "9ed9735b-66b5-44b3-9399-af938bfcd802" (UID: "9ed9735b-66b5-44b3-9399-af938bfcd802"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:16:15 crc kubenswrapper[4809]: I1124 07:16:15.532230 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ed9735b-66b5-44b3-9399-af938bfcd802-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 07:16:15 crc kubenswrapper[4809]: I1124 07:16:15.863669 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9" event={"ID":"9ed9735b-66b5-44b3-9399-af938bfcd802","Type":"ContainerDied","Data":"a6b5ab4a189341699779a665e29a04a55c439b9fa458868f313b42000935b54d"} Nov 24 07:16:15 crc kubenswrapper[4809]: I1124 07:16:15.863709 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6b5ab4a189341699779a665e29a04a55c439b9fa458868f313b42000935b54d" Nov 24 07:16:15 crc kubenswrapper[4809]: I1124 07:16:15.863812 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9" Nov 24 07:16:16 crc kubenswrapper[4809]: I1124 07:16:15.969307 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-wmc74"] Nov 24 07:16:16 crc kubenswrapper[4809]: E1124 07:16:15.969835 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ed9735b-66b5-44b3-9399-af938bfcd802" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 07:16:16 crc kubenswrapper[4809]: I1124 07:16:15.969856 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ed9735b-66b5-44b3-9399-af938bfcd802" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 07:16:16 crc kubenswrapper[4809]: I1124 07:16:15.970206 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ed9735b-66b5-44b3-9399-af938bfcd802" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 07:16:16 crc kubenswrapper[4809]: I1124 07:16:15.970913 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-wmc74" Nov 24 07:16:16 crc kubenswrapper[4809]: I1124 07:16:15.973095 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 07:16:16 crc kubenswrapper[4809]: I1124 07:16:15.973277 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lx7ll" Nov 24 07:16:16 crc kubenswrapper[4809]: I1124 07:16:15.975166 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 07:16:16 crc kubenswrapper[4809]: I1124 07:16:15.975345 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 07:16:16 crc kubenswrapper[4809]: I1124 07:16:15.986012 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-wmc74"] Nov 24 07:16:16 crc kubenswrapper[4809]: I1124 07:16:16.041492 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6r7b\" (UniqueName: \"kubernetes.io/projected/7883d395-4314-4e1c-9a89-89b949d1b7a6-kube-api-access-c6r7b\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-wmc74\" (UID: \"7883d395-4314-4e1c-9a89-89b949d1b7a6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-wmc74" Nov 24 07:16:16 crc kubenswrapper[4809]: I1124 07:16:16.041559 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7883d395-4314-4e1c-9a89-89b949d1b7a6-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-wmc74\" (UID: \"7883d395-4314-4e1c-9a89-89b949d1b7a6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-wmc74" Nov 24 07:16:16 crc kubenswrapper[4809]: I1124 07:16:16.043965 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7883d395-4314-4e1c-9a89-89b949d1b7a6-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-wmc74\" (UID: \"7883d395-4314-4e1c-9a89-89b949d1b7a6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-wmc74" Nov 24 07:16:16 crc kubenswrapper[4809]: I1124 07:16:16.145335 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7883d395-4314-4e1c-9a89-89b949d1b7a6-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-wmc74\" (UID: \"7883d395-4314-4e1c-9a89-89b949d1b7a6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-wmc74" Nov 24 07:16:16 crc kubenswrapper[4809]: I1124 07:16:16.145424 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7883d395-4314-4e1c-9a89-89b949d1b7a6-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-wmc74\" (UID: \"7883d395-4314-4e1c-9a89-89b949d1b7a6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-wmc74" Nov 24 07:16:16 crc kubenswrapper[4809]: I1124 07:16:16.145598 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6r7b\" (UniqueName: \"kubernetes.io/projected/7883d395-4314-4e1c-9a89-89b949d1b7a6-kube-api-access-c6r7b\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-wmc74\" (UID: \"7883d395-4314-4e1c-9a89-89b949d1b7a6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-wmc74" Nov 24 07:16:16 crc kubenswrapper[4809]: I1124 07:16:16.150761 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7883d395-4314-4e1c-9a89-89b949d1b7a6-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-wmc74\" (UID: \"7883d395-4314-4e1c-9a89-89b949d1b7a6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-wmc74" Nov 24 07:16:16 crc kubenswrapper[4809]: I1124 07:16:16.151572 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7883d395-4314-4e1c-9a89-89b949d1b7a6-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-wmc74\" (UID: \"7883d395-4314-4e1c-9a89-89b949d1b7a6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-wmc74" Nov 24 07:16:16 crc kubenswrapper[4809]: I1124 07:16:16.160519 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6r7b\" (UniqueName: \"kubernetes.io/projected/7883d395-4314-4e1c-9a89-89b949d1b7a6-kube-api-access-c6r7b\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-wmc74\" (UID: \"7883d395-4314-4e1c-9a89-89b949d1b7a6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-wmc74" Nov 24 07:16:16 crc kubenswrapper[4809]: I1124 07:16:16.327565 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-wmc74" Nov 24 07:16:16 crc kubenswrapper[4809]: I1124 07:16:16.908566 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-wmc74"] Nov 24 07:16:16 crc kubenswrapper[4809]: W1124 07:16:16.915247 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7883d395_4314_4e1c_9a89_89b949d1b7a6.slice/crio-edfd357546dd18414e3eb63eacd0a5307b74de048edb782baeaf42a88afde146 WatchSource:0}: Error finding container edfd357546dd18414e3eb63eacd0a5307b74de048edb782baeaf42a88afde146: Status 404 returned error can't find the container with id edfd357546dd18414e3eb63eacd0a5307b74de048edb782baeaf42a88afde146 Nov 24 07:16:17 crc kubenswrapper[4809]: I1124 07:16:17.884305 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-wmc74" event={"ID":"7883d395-4314-4e1c-9a89-89b949d1b7a6","Type":"ContainerStarted","Data":"0c293f7f78f20c2edd92b1eaed8e4da7d6927a2b7ad44e571e13eac852bf01af"} Nov 24 07:16:17 crc kubenswrapper[4809]: I1124 07:16:17.884650 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-wmc74" event={"ID":"7883d395-4314-4e1c-9a89-89b949d1b7a6","Type":"ContainerStarted","Data":"edfd357546dd18414e3eb63eacd0a5307b74de048edb782baeaf42a88afde146"} Nov 24 07:16:18 crc kubenswrapper[4809]: I1124 07:16:18.043857 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:16:18 crc kubenswrapper[4809]: I1124 07:16:18.043914 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:16:20 crc kubenswrapper[4809]: I1124 07:16:20.913832 4809 generic.go:334] "Generic (PLEG): container finished" podID="7883d395-4314-4e1c-9a89-89b949d1b7a6" containerID="0c293f7f78f20c2edd92b1eaed8e4da7d6927a2b7ad44e571e13eac852bf01af" exitCode=0 Nov 24 07:16:20 crc kubenswrapper[4809]: I1124 07:16:20.913924 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-wmc74" event={"ID":"7883d395-4314-4e1c-9a89-89b949d1b7a6","Type":"ContainerDied","Data":"0c293f7f78f20c2edd92b1eaed8e4da7d6927a2b7ad44e571e13eac852bf01af"} Nov 24 07:16:22 crc kubenswrapper[4809]: I1124 07:16:22.351274 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-wmc74" Nov 24 07:16:22 crc kubenswrapper[4809]: I1124 07:16:22.461815 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6r7b\" (UniqueName: \"kubernetes.io/projected/7883d395-4314-4e1c-9a89-89b949d1b7a6-kube-api-access-c6r7b\") pod \"7883d395-4314-4e1c-9a89-89b949d1b7a6\" (UID: \"7883d395-4314-4e1c-9a89-89b949d1b7a6\") " Nov 24 07:16:22 crc kubenswrapper[4809]: I1124 07:16:22.461905 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7883d395-4314-4e1c-9a89-89b949d1b7a6-ssh-key\") pod \"7883d395-4314-4e1c-9a89-89b949d1b7a6\" (UID: \"7883d395-4314-4e1c-9a89-89b949d1b7a6\") " Nov 24 07:16:22 crc kubenswrapper[4809]: I1124 07:16:22.461990 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7883d395-4314-4e1c-9a89-89b949d1b7a6-inventory\") pod \"7883d395-4314-4e1c-9a89-89b949d1b7a6\" (UID: \"7883d395-4314-4e1c-9a89-89b949d1b7a6\") " Nov 24 07:16:22 crc kubenswrapper[4809]: I1124 07:16:22.467359 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7883d395-4314-4e1c-9a89-89b949d1b7a6-kube-api-access-c6r7b" (OuterVolumeSpecName: "kube-api-access-c6r7b") pod "7883d395-4314-4e1c-9a89-89b949d1b7a6" (UID: "7883d395-4314-4e1c-9a89-89b949d1b7a6"). InnerVolumeSpecName "kube-api-access-c6r7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:16:22 crc kubenswrapper[4809]: I1124 07:16:22.491223 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7883d395-4314-4e1c-9a89-89b949d1b7a6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7883d395-4314-4e1c-9a89-89b949d1b7a6" (UID: "7883d395-4314-4e1c-9a89-89b949d1b7a6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:16:22 crc kubenswrapper[4809]: I1124 07:16:22.493155 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7883d395-4314-4e1c-9a89-89b949d1b7a6-inventory" (OuterVolumeSpecName: "inventory") pod "7883d395-4314-4e1c-9a89-89b949d1b7a6" (UID: "7883d395-4314-4e1c-9a89-89b949d1b7a6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:16:22 crc kubenswrapper[4809]: I1124 07:16:22.563615 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6r7b\" (UniqueName: \"kubernetes.io/projected/7883d395-4314-4e1c-9a89-89b949d1b7a6-kube-api-access-c6r7b\") on node \"crc\" DevicePath \"\"" Nov 24 07:16:22 crc kubenswrapper[4809]: I1124 07:16:22.563649 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7883d395-4314-4e1c-9a89-89b949d1b7a6-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 07:16:22 crc kubenswrapper[4809]: I1124 07:16:22.563658 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7883d395-4314-4e1c-9a89-89b949d1b7a6-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 07:16:22 crc kubenswrapper[4809]: I1124 07:16:22.945764 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-wmc74" event={"ID":"7883d395-4314-4e1c-9a89-89b949d1b7a6","Type":"ContainerDied","Data":"edfd357546dd18414e3eb63eacd0a5307b74de048edb782baeaf42a88afde146"} Nov 24 07:16:22 crc kubenswrapper[4809]: I1124 07:16:22.945822 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="edfd357546dd18414e3eb63eacd0a5307b74de048edb782baeaf42a88afde146" Nov 24 07:16:22 crc kubenswrapper[4809]: I1124 07:16:22.945821 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-wmc74" Nov 24 07:16:23 crc kubenswrapper[4809]: I1124 07:16:23.027235 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28"] Nov 24 07:16:23 crc kubenswrapper[4809]: E1124 07:16:23.027769 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7883d395-4314-4e1c-9a89-89b949d1b7a6" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 24 07:16:23 crc kubenswrapper[4809]: I1124 07:16:23.027793 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7883d395-4314-4e1c-9a89-89b949d1b7a6" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 24 07:16:23 crc kubenswrapper[4809]: I1124 07:16:23.028082 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7883d395-4314-4e1c-9a89-89b949d1b7a6" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 24 07:16:23 crc kubenswrapper[4809]: I1124 07:16:23.028913 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28" Nov 24 07:16:23 crc kubenswrapper[4809]: I1124 07:16:23.034969 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28"] Nov 24 07:16:23 crc kubenswrapper[4809]: I1124 07:16:23.037466 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 07:16:23 crc kubenswrapper[4809]: I1124 07:16:23.037596 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 07:16:23 crc kubenswrapper[4809]: I1124 07:16:23.038000 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 07:16:23 crc kubenswrapper[4809]: I1124 07:16:23.038380 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lx7ll" Nov 24 07:16:23 crc kubenswrapper[4809]: I1124 07:16:23.079039 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bad47712-c44f-4799-bfe2-6c95ea3b513f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28\" (UID: \"bad47712-c44f-4799-bfe2-6c95ea3b513f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28" Nov 24 07:16:23 crc kubenswrapper[4809]: I1124 07:16:23.079185 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bad47712-c44f-4799-bfe2-6c95ea3b513f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28\" (UID: \"bad47712-c44f-4799-bfe2-6c95ea3b513f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28" Nov 24 07:16:23 crc kubenswrapper[4809]: I1124 07:16:23.079231 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad47712-c44f-4799-bfe2-6c95ea3b513f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28\" (UID: \"bad47712-c44f-4799-bfe2-6c95ea3b513f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28" Nov 24 07:16:23 crc kubenswrapper[4809]: I1124 07:16:23.079527 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvz8w\" (UniqueName: \"kubernetes.io/projected/bad47712-c44f-4799-bfe2-6c95ea3b513f-kube-api-access-bvz8w\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28\" (UID: \"bad47712-c44f-4799-bfe2-6c95ea3b513f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28" Nov 24 07:16:23 crc kubenswrapper[4809]: I1124 07:16:23.180647 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bad47712-c44f-4799-bfe2-6c95ea3b513f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28\" (UID: \"bad47712-c44f-4799-bfe2-6c95ea3b513f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28" Nov 24 07:16:23 crc kubenswrapper[4809]: I1124 07:16:23.180712 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bad47712-c44f-4799-bfe2-6c95ea3b513f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28\" (UID: \"bad47712-c44f-4799-bfe2-6c95ea3b513f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28" Nov 24 07:16:23 crc kubenswrapper[4809]: I1124 07:16:23.180738 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad47712-c44f-4799-bfe2-6c95ea3b513f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28\" (UID: \"bad47712-c44f-4799-bfe2-6c95ea3b513f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28" Nov 24 07:16:23 crc kubenswrapper[4809]: I1124 07:16:23.180796 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvz8w\" (UniqueName: \"kubernetes.io/projected/bad47712-c44f-4799-bfe2-6c95ea3b513f-kube-api-access-bvz8w\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28\" (UID: \"bad47712-c44f-4799-bfe2-6c95ea3b513f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28" Nov 24 07:16:23 crc kubenswrapper[4809]: I1124 07:16:23.186204 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bad47712-c44f-4799-bfe2-6c95ea3b513f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28\" (UID: \"bad47712-c44f-4799-bfe2-6c95ea3b513f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28" Nov 24 07:16:23 crc kubenswrapper[4809]: I1124 07:16:23.190465 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bad47712-c44f-4799-bfe2-6c95ea3b513f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28\" (UID: \"bad47712-c44f-4799-bfe2-6c95ea3b513f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28" Nov 24 07:16:23 crc kubenswrapper[4809]: I1124 07:16:23.194441 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad47712-c44f-4799-bfe2-6c95ea3b513f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28\" (UID: \"bad47712-c44f-4799-bfe2-6c95ea3b513f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28" Nov 24 07:16:23 crc kubenswrapper[4809]: I1124 07:16:23.210323 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvz8w\" (UniqueName: \"kubernetes.io/projected/bad47712-c44f-4799-bfe2-6c95ea3b513f-kube-api-access-bvz8w\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28\" (UID: \"bad47712-c44f-4799-bfe2-6c95ea3b513f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28" Nov 24 07:16:23 crc kubenswrapper[4809]: I1124 07:16:23.353926 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28" Nov 24 07:16:24 crc kubenswrapper[4809]: I1124 07:16:24.021593 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28"] Nov 24 07:16:24 crc kubenswrapper[4809]: W1124 07:16:24.029154 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbad47712_c44f_4799_bfe2_6c95ea3b513f.slice/crio-74163b98736e7ab8efa930e3174349e09348b832cbca5dfa14f91f76ff1925d3 WatchSource:0}: Error finding container 74163b98736e7ab8efa930e3174349e09348b832cbca5dfa14f91f76ff1925d3: Status 404 returned error can't find the container with id 74163b98736e7ab8efa930e3174349e09348b832cbca5dfa14f91f76ff1925d3 Nov 24 07:16:24 crc kubenswrapper[4809]: I1124 07:16:24.966164 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28" event={"ID":"bad47712-c44f-4799-bfe2-6c95ea3b513f","Type":"ContainerStarted","Data":"f38b4bf098bbc323eaf151c0fa60f6c3b147b8d534efe7dfbc08d219cec584e1"} Nov 24 07:16:24 crc kubenswrapper[4809]: I1124 07:16:24.966575 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28" event={"ID":"bad47712-c44f-4799-bfe2-6c95ea3b513f","Type":"ContainerStarted","Data":"74163b98736e7ab8efa930e3174349e09348b832cbca5dfa14f91f76ff1925d3"} Nov 24 07:16:24 crc kubenswrapper[4809]: I1124 07:16:24.989352 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28" podStartSLOduration=2.39611052 podStartE2EDuration="2.989330887s" podCreationTimestamp="2025-11-24 07:16:22 +0000 UTC" firstStartedPulling="2025-11-24 07:16:24.034401794 +0000 UTC m=+1283.934993499" lastFinishedPulling="2025-11-24 07:16:24.627622161 +0000 UTC m=+1284.528213866" observedRunningTime="2025-11-24 07:16:24.982356131 +0000 UTC m=+1284.882947866" watchObservedRunningTime="2025-11-24 07:16:24.989330887 +0000 UTC m=+1284.889922592" Nov 24 07:16:48 crc kubenswrapper[4809]: I1124 07:16:48.043627 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:16:48 crc kubenswrapper[4809]: I1124 07:16:48.044149 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:16:48 crc kubenswrapper[4809]: I1124 07:16:48.044210 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 07:16:48 crc kubenswrapper[4809]: I1124 07:16:48.045158 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"11616e3dc08f43f601be45c4cf1cbd156c734d2c4a8bf401dd3c381f1cf4aefb"} pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:16:48 crc kubenswrapper[4809]: I1124 07:16:48.045246 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" containerID="cri-o://11616e3dc08f43f601be45c4cf1cbd156c734d2c4a8bf401dd3c381f1cf4aefb" gracePeriod=600 Nov 24 07:16:48 crc kubenswrapper[4809]: I1124 07:16:48.211105 4809 generic.go:334] "Generic (PLEG): container finished" podID="be8050d5-1aea-453e-a06e-87702b0e856b" containerID="11616e3dc08f43f601be45c4cf1cbd156c734d2c4a8bf401dd3c381f1cf4aefb" exitCode=0 Nov 24 07:16:48 crc kubenswrapper[4809]: I1124 07:16:48.211161 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerDied","Data":"11616e3dc08f43f601be45c4cf1cbd156c734d2c4a8bf401dd3c381f1cf4aefb"} Nov 24 07:16:48 crc kubenswrapper[4809]: I1124 07:16:48.211200 4809 scope.go:117] "RemoveContainer" containerID="37feb2230d7f506014799017af7463c0d7873535f5cdb2251d0334f2328d10b8" Nov 24 07:16:49 crc kubenswrapper[4809]: I1124 07:16:49.226448 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerStarted","Data":"e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e"} Nov 24 07:17:07 crc kubenswrapper[4809]: I1124 07:17:07.248165 4809 scope.go:117] "RemoveContainer" containerID="3618f80dfb48da6ff14b44796854bdfabe49327124842a2b566bffb6b744df7e" Nov 24 07:17:07 crc kubenswrapper[4809]: I1124 07:17:07.288666 4809 scope.go:117] "RemoveContainer" containerID="e2e3af7c670851a4e17e2da182ad3af767d4ca7e2fd95e7e821fbe3d4b4a1b00" Nov 24 07:18:07 crc kubenswrapper[4809]: I1124 07:18:07.428538 4809 scope.go:117] "RemoveContainer" containerID="4436bea8e3c221175588f21ce0a5d2e2af759b714bd7d68636b3cf70e819c4c4" Nov 24 07:18:07 crc kubenswrapper[4809]: I1124 07:18:07.466461 4809 scope.go:117] "RemoveContainer" containerID="b70526d14ae2999033922d90886b26dd0c6962c84bdd995e51968f2ff939118c" Nov 24 07:18:07 crc kubenswrapper[4809]: I1124 07:18:07.576245 4809 scope.go:117] "RemoveContainer" containerID="9847ef762be1e8b052fdd7e6a83f18c7bc41cb4c56be305015af58ce1c31776e" Nov 24 07:18:22 crc kubenswrapper[4809]: I1124 07:18:22.753821 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2bd2q"] Nov 24 07:18:22 crc kubenswrapper[4809]: I1124 07:18:22.760659 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2bd2q" Nov 24 07:18:22 crc kubenswrapper[4809]: I1124 07:18:22.770522 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2bd2q"] Nov 24 07:18:22 crc kubenswrapper[4809]: I1124 07:18:22.771923 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x92ns\" (UniqueName: \"kubernetes.io/projected/5d435941-9f2b-44e0-b18f-33d57d505e39-kube-api-access-x92ns\") pod \"redhat-marketplace-2bd2q\" (UID: \"5d435941-9f2b-44e0-b18f-33d57d505e39\") " pod="openshift-marketplace/redhat-marketplace-2bd2q" Nov 24 07:18:22 crc kubenswrapper[4809]: I1124 07:18:22.772240 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d435941-9f2b-44e0-b18f-33d57d505e39-catalog-content\") pod \"redhat-marketplace-2bd2q\" (UID: \"5d435941-9f2b-44e0-b18f-33d57d505e39\") " pod="openshift-marketplace/redhat-marketplace-2bd2q" Nov 24 07:18:22 crc kubenswrapper[4809]: I1124 07:18:22.772439 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d435941-9f2b-44e0-b18f-33d57d505e39-utilities\") pod \"redhat-marketplace-2bd2q\" (UID: \"5d435941-9f2b-44e0-b18f-33d57d505e39\") " pod="openshift-marketplace/redhat-marketplace-2bd2q" Nov 24 07:18:22 crc kubenswrapper[4809]: I1124 07:18:22.873353 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x92ns\" (UniqueName: \"kubernetes.io/projected/5d435941-9f2b-44e0-b18f-33d57d505e39-kube-api-access-x92ns\") pod \"redhat-marketplace-2bd2q\" (UID: \"5d435941-9f2b-44e0-b18f-33d57d505e39\") " pod="openshift-marketplace/redhat-marketplace-2bd2q" Nov 24 07:18:22 crc kubenswrapper[4809]: I1124 07:18:22.873397 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d435941-9f2b-44e0-b18f-33d57d505e39-catalog-content\") pod \"redhat-marketplace-2bd2q\" (UID: \"5d435941-9f2b-44e0-b18f-33d57d505e39\") " pod="openshift-marketplace/redhat-marketplace-2bd2q" Nov 24 07:18:22 crc kubenswrapper[4809]: I1124 07:18:22.873429 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d435941-9f2b-44e0-b18f-33d57d505e39-utilities\") pod \"redhat-marketplace-2bd2q\" (UID: \"5d435941-9f2b-44e0-b18f-33d57d505e39\") " pod="openshift-marketplace/redhat-marketplace-2bd2q" Nov 24 07:18:22 crc kubenswrapper[4809]: I1124 07:18:22.873836 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d435941-9f2b-44e0-b18f-33d57d505e39-utilities\") pod \"redhat-marketplace-2bd2q\" (UID: \"5d435941-9f2b-44e0-b18f-33d57d505e39\") " pod="openshift-marketplace/redhat-marketplace-2bd2q" Nov 24 07:18:22 crc kubenswrapper[4809]: I1124 07:18:22.874103 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d435941-9f2b-44e0-b18f-33d57d505e39-catalog-content\") pod \"redhat-marketplace-2bd2q\" (UID: \"5d435941-9f2b-44e0-b18f-33d57d505e39\") " pod="openshift-marketplace/redhat-marketplace-2bd2q" Nov 24 07:18:22 crc kubenswrapper[4809]: I1124 07:18:22.900689 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x92ns\" (UniqueName: \"kubernetes.io/projected/5d435941-9f2b-44e0-b18f-33d57d505e39-kube-api-access-x92ns\") pod \"redhat-marketplace-2bd2q\" (UID: \"5d435941-9f2b-44e0-b18f-33d57d505e39\") " pod="openshift-marketplace/redhat-marketplace-2bd2q" Nov 24 07:18:23 crc kubenswrapper[4809]: I1124 07:18:23.144342 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2bd2q" Nov 24 07:18:23 crc kubenswrapper[4809]: I1124 07:18:23.631664 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2bd2q"] Nov 24 07:18:23 crc kubenswrapper[4809]: W1124 07:18:23.636540 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d435941_9f2b_44e0_b18f_33d57d505e39.slice/crio-f877900ebe17493ea90839bf82e3ce16c4b0a4c36a610a4266ac0e77227d38c0 WatchSource:0}: Error finding container f877900ebe17493ea90839bf82e3ce16c4b0a4c36a610a4266ac0e77227d38c0: Status 404 returned error can't find the container with id f877900ebe17493ea90839bf82e3ce16c4b0a4c36a610a4266ac0e77227d38c0 Nov 24 07:18:24 crc kubenswrapper[4809]: I1124 07:18:24.243199 4809 generic.go:334] "Generic (PLEG): container finished" podID="5d435941-9f2b-44e0-b18f-33d57d505e39" containerID="0d8dab848781672d64d019ef3344eb7a9f8a56864bc62902538c2042423b6814" exitCode=0 Nov 24 07:18:24 crc kubenswrapper[4809]: I1124 07:18:24.243267 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2bd2q" event={"ID":"5d435941-9f2b-44e0-b18f-33d57d505e39","Type":"ContainerDied","Data":"0d8dab848781672d64d019ef3344eb7a9f8a56864bc62902538c2042423b6814"} Nov 24 07:18:24 crc kubenswrapper[4809]: I1124 07:18:24.243564 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2bd2q" event={"ID":"5d435941-9f2b-44e0-b18f-33d57d505e39","Type":"ContainerStarted","Data":"f877900ebe17493ea90839bf82e3ce16c4b0a4c36a610a4266ac0e77227d38c0"} Nov 24 07:18:25 crc kubenswrapper[4809]: I1124 07:18:25.256308 4809 generic.go:334] "Generic (PLEG): container finished" podID="5d435941-9f2b-44e0-b18f-33d57d505e39" containerID="ec01058d9a3b8065ee1b68981ccef94ea9a576ffc71fb35dbefd5856f6fc5390" exitCode=0 Nov 24 07:18:25 crc kubenswrapper[4809]: I1124 07:18:25.256354 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2bd2q" event={"ID":"5d435941-9f2b-44e0-b18f-33d57d505e39","Type":"ContainerDied","Data":"ec01058d9a3b8065ee1b68981ccef94ea9a576ffc71fb35dbefd5856f6fc5390"} Nov 24 07:18:26 crc kubenswrapper[4809]: I1124 07:18:26.268638 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2bd2q" event={"ID":"5d435941-9f2b-44e0-b18f-33d57d505e39","Type":"ContainerStarted","Data":"56d1aa6ea34dc434533661ea3b0a43afe435084550c02911d424ce23469d23e8"} Nov 24 07:18:26 crc kubenswrapper[4809]: I1124 07:18:26.291161 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2bd2q" podStartSLOduration=2.751257801 podStartE2EDuration="4.29114415s" podCreationTimestamp="2025-11-24 07:18:22 +0000 UTC" firstStartedPulling="2025-11-24 07:18:24.246591288 +0000 UTC m=+1404.147183033" lastFinishedPulling="2025-11-24 07:18:25.786477667 +0000 UTC m=+1405.687069382" observedRunningTime="2025-11-24 07:18:26.287276456 +0000 UTC m=+1406.187868181" watchObservedRunningTime="2025-11-24 07:18:26.29114415 +0000 UTC m=+1406.191735855" Nov 24 07:18:29 crc kubenswrapper[4809]: I1124 07:18:29.134377 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rjlcg"] Nov 24 07:18:29 crc kubenswrapper[4809]: I1124 07:18:29.140123 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rjlcg" Nov 24 07:18:29 crc kubenswrapper[4809]: I1124 07:18:29.153321 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rjlcg"] Nov 24 07:18:29 crc kubenswrapper[4809]: I1124 07:18:29.294438 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21b5fc68-cb78-4a1f-b924-25720d84deb4-utilities\") pod \"community-operators-rjlcg\" (UID: \"21b5fc68-cb78-4a1f-b924-25720d84deb4\") " pod="openshift-marketplace/community-operators-rjlcg" Nov 24 07:18:29 crc kubenswrapper[4809]: I1124 07:18:29.294560 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k97r\" (UniqueName: \"kubernetes.io/projected/21b5fc68-cb78-4a1f-b924-25720d84deb4-kube-api-access-2k97r\") pod \"community-operators-rjlcg\" (UID: \"21b5fc68-cb78-4a1f-b924-25720d84deb4\") " pod="openshift-marketplace/community-operators-rjlcg" Nov 24 07:18:29 crc kubenswrapper[4809]: I1124 07:18:29.294609 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21b5fc68-cb78-4a1f-b924-25720d84deb4-catalog-content\") pod \"community-operators-rjlcg\" (UID: \"21b5fc68-cb78-4a1f-b924-25720d84deb4\") " pod="openshift-marketplace/community-operators-rjlcg" Nov 24 07:18:29 crc kubenswrapper[4809]: I1124 07:18:29.396179 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k97r\" (UniqueName: \"kubernetes.io/projected/21b5fc68-cb78-4a1f-b924-25720d84deb4-kube-api-access-2k97r\") pod \"community-operators-rjlcg\" (UID: \"21b5fc68-cb78-4a1f-b924-25720d84deb4\") " pod="openshift-marketplace/community-operators-rjlcg" Nov 24 07:18:29 crc kubenswrapper[4809]: I1124 07:18:29.396477 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21b5fc68-cb78-4a1f-b924-25720d84deb4-catalog-content\") pod \"community-operators-rjlcg\" (UID: \"21b5fc68-cb78-4a1f-b924-25720d84deb4\") " pod="openshift-marketplace/community-operators-rjlcg" Nov 24 07:18:29 crc kubenswrapper[4809]: I1124 07:18:29.396587 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21b5fc68-cb78-4a1f-b924-25720d84deb4-utilities\") pod \"community-operators-rjlcg\" (UID: \"21b5fc68-cb78-4a1f-b924-25720d84deb4\") " pod="openshift-marketplace/community-operators-rjlcg" Nov 24 07:18:29 crc kubenswrapper[4809]: I1124 07:18:29.397121 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21b5fc68-cb78-4a1f-b924-25720d84deb4-utilities\") pod \"community-operators-rjlcg\" (UID: \"21b5fc68-cb78-4a1f-b924-25720d84deb4\") " pod="openshift-marketplace/community-operators-rjlcg" Nov 24 07:18:29 crc kubenswrapper[4809]: I1124 07:18:29.397467 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21b5fc68-cb78-4a1f-b924-25720d84deb4-catalog-content\") pod \"community-operators-rjlcg\" (UID: \"21b5fc68-cb78-4a1f-b924-25720d84deb4\") " pod="openshift-marketplace/community-operators-rjlcg" Nov 24 07:18:29 crc kubenswrapper[4809]: I1124 07:18:29.416484 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k97r\" (UniqueName: \"kubernetes.io/projected/21b5fc68-cb78-4a1f-b924-25720d84deb4-kube-api-access-2k97r\") pod \"community-operators-rjlcg\" (UID: \"21b5fc68-cb78-4a1f-b924-25720d84deb4\") " pod="openshift-marketplace/community-operators-rjlcg" Nov 24 07:18:29 crc kubenswrapper[4809]: I1124 07:18:29.506123 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rjlcg" Nov 24 07:18:30 crc kubenswrapper[4809]: I1124 07:18:30.033562 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rjlcg"] Nov 24 07:18:30 crc kubenswrapper[4809]: W1124 07:18:30.036921 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21b5fc68_cb78_4a1f_b924_25720d84deb4.slice/crio-1f72d6e6e439a59c1ae6b2b5fdddc8532eab9f7ff15572e3f8b3a94478ffaafa WatchSource:0}: Error finding container 1f72d6e6e439a59c1ae6b2b5fdddc8532eab9f7ff15572e3f8b3a94478ffaafa: Status 404 returned error can't find the container with id 1f72d6e6e439a59c1ae6b2b5fdddc8532eab9f7ff15572e3f8b3a94478ffaafa Nov 24 07:18:30 crc kubenswrapper[4809]: I1124 07:18:30.325664 4809 generic.go:334] "Generic (PLEG): container finished" podID="21b5fc68-cb78-4a1f-b924-25720d84deb4" containerID="09d2e695d2631a38547b6b756c78ebcfb9b7a4db9c9a376ba6706803a670361c" exitCode=0 Nov 24 07:18:30 crc kubenswrapper[4809]: I1124 07:18:30.325704 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjlcg" event={"ID":"21b5fc68-cb78-4a1f-b924-25720d84deb4","Type":"ContainerDied","Data":"09d2e695d2631a38547b6b756c78ebcfb9b7a4db9c9a376ba6706803a670361c"} Nov 24 07:18:30 crc kubenswrapper[4809]: I1124 07:18:30.325732 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjlcg" event={"ID":"21b5fc68-cb78-4a1f-b924-25720d84deb4","Type":"ContainerStarted","Data":"1f72d6e6e439a59c1ae6b2b5fdddc8532eab9f7ff15572e3f8b3a94478ffaafa"} Nov 24 07:18:31 crc kubenswrapper[4809]: I1124 07:18:31.345690 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjlcg" event={"ID":"21b5fc68-cb78-4a1f-b924-25720d84deb4","Type":"ContainerStarted","Data":"5a797d1d3358a0dc27ad771e4795628b9884fad19c9f6508b8e4cae9a716eb62"} Nov 24 07:18:32 crc kubenswrapper[4809]: I1124 07:18:32.356210 4809 generic.go:334] "Generic (PLEG): container finished" podID="21b5fc68-cb78-4a1f-b924-25720d84deb4" containerID="5a797d1d3358a0dc27ad771e4795628b9884fad19c9f6508b8e4cae9a716eb62" exitCode=0 Nov 24 07:18:32 crc kubenswrapper[4809]: I1124 07:18:32.356318 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjlcg" event={"ID":"21b5fc68-cb78-4a1f-b924-25720d84deb4","Type":"ContainerDied","Data":"5a797d1d3358a0dc27ad771e4795628b9884fad19c9f6508b8e4cae9a716eb62"} Nov 24 07:18:33 crc kubenswrapper[4809]: I1124 07:18:33.144585 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2bd2q" Nov 24 07:18:33 crc kubenswrapper[4809]: I1124 07:18:33.145373 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2bd2q" Nov 24 07:18:33 crc kubenswrapper[4809]: I1124 07:18:33.216994 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2bd2q" Nov 24 07:18:33 crc kubenswrapper[4809]: I1124 07:18:33.382429 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjlcg" event={"ID":"21b5fc68-cb78-4a1f-b924-25720d84deb4","Type":"ContainerStarted","Data":"526e9f59e535aa0bb976427c59b6c8ed02e433d6a478d9cbfc71b7e0941cb925"} Nov 24 07:18:33 crc kubenswrapper[4809]: I1124 07:18:33.415946 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rjlcg" podStartSLOduration=1.962837295 podStartE2EDuration="4.415922394s" podCreationTimestamp="2025-11-24 07:18:29 +0000 UTC" firstStartedPulling="2025-11-24 07:18:30.328786419 +0000 UTC m=+1410.229378124" lastFinishedPulling="2025-11-24 07:18:32.781871528 +0000 UTC m=+1412.682463223" observedRunningTime="2025-11-24 07:18:33.403896804 +0000 UTC m=+1413.304488519" watchObservedRunningTime="2025-11-24 07:18:33.415922394 +0000 UTC m=+1413.316514109" Nov 24 07:18:33 crc kubenswrapper[4809]: I1124 07:18:33.455185 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2bd2q" Nov 24 07:18:35 crc kubenswrapper[4809]: I1124 07:18:35.526048 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2bd2q"] Nov 24 07:18:35 crc kubenswrapper[4809]: I1124 07:18:35.526482 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2bd2q" podUID="5d435941-9f2b-44e0-b18f-33d57d505e39" containerName="registry-server" containerID="cri-o://56d1aa6ea34dc434533661ea3b0a43afe435084550c02911d424ce23469d23e8" gracePeriod=2 Nov 24 07:18:35 crc kubenswrapper[4809]: I1124 07:18:35.970301 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2bd2q" Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.129958 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d435941-9f2b-44e0-b18f-33d57d505e39-catalog-content\") pod \"5d435941-9f2b-44e0-b18f-33d57d505e39\" (UID: \"5d435941-9f2b-44e0-b18f-33d57d505e39\") " Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.130121 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x92ns\" (UniqueName: \"kubernetes.io/projected/5d435941-9f2b-44e0-b18f-33d57d505e39-kube-api-access-x92ns\") pod \"5d435941-9f2b-44e0-b18f-33d57d505e39\" (UID: \"5d435941-9f2b-44e0-b18f-33d57d505e39\") " Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.130214 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d435941-9f2b-44e0-b18f-33d57d505e39-utilities\") pod \"5d435941-9f2b-44e0-b18f-33d57d505e39\" (UID: \"5d435941-9f2b-44e0-b18f-33d57d505e39\") " Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.131416 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d435941-9f2b-44e0-b18f-33d57d505e39-utilities" (OuterVolumeSpecName: "utilities") pod "5d435941-9f2b-44e0-b18f-33d57d505e39" (UID: "5d435941-9f2b-44e0-b18f-33d57d505e39"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.137223 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d435941-9f2b-44e0-b18f-33d57d505e39-kube-api-access-x92ns" (OuterVolumeSpecName: "kube-api-access-x92ns") pod "5d435941-9f2b-44e0-b18f-33d57d505e39" (UID: "5d435941-9f2b-44e0-b18f-33d57d505e39"). InnerVolumeSpecName "kube-api-access-x92ns". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.152828 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d435941-9f2b-44e0-b18f-33d57d505e39-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5d435941-9f2b-44e0-b18f-33d57d505e39" (UID: "5d435941-9f2b-44e0-b18f-33d57d505e39"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.232341 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x92ns\" (UniqueName: \"kubernetes.io/projected/5d435941-9f2b-44e0-b18f-33d57d505e39-kube-api-access-x92ns\") on node \"crc\" DevicePath \"\"" Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.232388 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d435941-9f2b-44e0-b18f-33d57d505e39-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.232407 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d435941-9f2b-44e0-b18f-33d57d505e39-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.416830 4809 generic.go:334] "Generic (PLEG): container finished" podID="5d435941-9f2b-44e0-b18f-33d57d505e39" containerID="56d1aa6ea34dc434533661ea3b0a43afe435084550c02911d424ce23469d23e8" exitCode=0 Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.416890 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2bd2q" event={"ID":"5d435941-9f2b-44e0-b18f-33d57d505e39","Type":"ContainerDied","Data":"56d1aa6ea34dc434533661ea3b0a43afe435084550c02911d424ce23469d23e8"} Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.416919 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2bd2q" Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.416949 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2bd2q" event={"ID":"5d435941-9f2b-44e0-b18f-33d57d505e39","Type":"ContainerDied","Data":"f877900ebe17493ea90839bf82e3ce16c4b0a4c36a610a4266ac0e77227d38c0"} Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.417028 4809 scope.go:117] "RemoveContainer" containerID="56d1aa6ea34dc434533661ea3b0a43afe435084550c02911d424ce23469d23e8" Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.440858 4809 scope.go:117] "RemoveContainer" containerID="ec01058d9a3b8065ee1b68981ccef94ea9a576ffc71fb35dbefd5856f6fc5390" Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.466576 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2bd2q"] Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.478538 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2bd2q"] Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.496606 4809 scope.go:117] "RemoveContainer" containerID="0d8dab848781672d64d019ef3344eb7a9f8a56864bc62902538c2042423b6814" Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.520197 4809 scope.go:117] "RemoveContainer" containerID="56d1aa6ea34dc434533661ea3b0a43afe435084550c02911d424ce23469d23e8" Nov 24 07:18:36 crc kubenswrapper[4809]: E1124 07:18:36.520698 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56d1aa6ea34dc434533661ea3b0a43afe435084550c02911d424ce23469d23e8\": container with ID starting with 56d1aa6ea34dc434533661ea3b0a43afe435084550c02911d424ce23469d23e8 not found: ID does not exist" containerID="56d1aa6ea34dc434533661ea3b0a43afe435084550c02911d424ce23469d23e8" Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.520744 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56d1aa6ea34dc434533661ea3b0a43afe435084550c02911d424ce23469d23e8"} err="failed to get container status \"56d1aa6ea34dc434533661ea3b0a43afe435084550c02911d424ce23469d23e8\": rpc error: code = NotFound desc = could not find container \"56d1aa6ea34dc434533661ea3b0a43afe435084550c02911d424ce23469d23e8\": container with ID starting with 56d1aa6ea34dc434533661ea3b0a43afe435084550c02911d424ce23469d23e8 not found: ID does not exist" Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.520774 4809 scope.go:117] "RemoveContainer" containerID="ec01058d9a3b8065ee1b68981ccef94ea9a576ffc71fb35dbefd5856f6fc5390" Nov 24 07:18:36 crc kubenswrapper[4809]: E1124 07:18:36.521178 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec01058d9a3b8065ee1b68981ccef94ea9a576ffc71fb35dbefd5856f6fc5390\": container with ID starting with ec01058d9a3b8065ee1b68981ccef94ea9a576ffc71fb35dbefd5856f6fc5390 not found: ID does not exist" containerID="ec01058d9a3b8065ee1b68981ccef94ea9a576ffc71fb35dbefd5856f6fc5390" Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.521212 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec01058d9a3b8065ee1b68981ccef94ea9a576ffc71fb35dbefd5856f6fc5390"} err="failed to get container status \"ec01058d9a3b8065ee1b68981ccef94ea9a576ffc71fb35dbefd5856f6fc5390\": rpc error: code = NotFound desc = could not find container \"ec01058d9a3b8065ee1b68981ccef94ea9a576ffc71fb35dbefd5856f6fc5390\": container with ID starting with ec01058d9a3b8065ee1b68981ccef94ea9a576ffc71fb35dbefd5856f6fc5390 not found: ID does not exist" Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.521230 4809 scope.go:117] "RemoveContainer" containerID="0d8dab848781672d64d019ef3344eb7a9f8a56864bc62902538c2042423b6814" Nov 24 07:18:36 crc kubenswrapper[4809]: E1124 07:18:36.521624 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d8dab848781672d64d019ef3344eb7a9f8a56864bc62902538c2042423b6814\": container with ID starting with 0d8dab848781672d64d019ef3344eb7a9f8a56864bc62902538c2042423b6814 not found: ID does not exist" containerID="0d8dab848781672d64d019ef3344eb7a9f8a56864bc62902538c2042423b6814" Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.521665 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d8dab848781672d64d019ef3344eb7a9f8a56864bc62902538c2042423b6814"} err="failed to get container status \"0d8dab848781672d64d019ef3344eb7a9f8a56864bc62902538c2042423b6814\": rpc error: code = NotFound desc = could not find container \"0d8dab848781672d64d019ef3344eb7a9f8a56864bc62902538c2042423b6814\": container with ID starting with 0d8dab848781672d64d019ef3344eb7a9f8a56864bc62902538c2042423b6814 not found: ID does not exist" Nov 24 07:18:36 crc kubenswrapper[4809]: I1124 07:18:36.903789 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d435941-9f2b-44e0-b18f-33d57d505e39" path="/var/lib/kubelet/pods/5d435941-9f2b-44e0-b18f-33d57d505e39/volumes" Nov 24 07:18:39 crc kubenswrapper[4809]: I1124 07:18:39.506634 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rjlcg" Nov 24 07:18:39 crc kubenswrapper[4809]: I1124 07:18:39.507044 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rjlcg" Nov 24 07:18:39 crc kubenswrapper[4809]: I1124 07:18:39.538713 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k687t"] Nov 24 07:18:39 crc kubenswrapper[4809]: E1124 07:18:39.539183 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d435941-9f2b-44e0-b18f-33d57d505e39" containerName="registry-server" Nov 24 07:18:39 crc kubenswrapper[4809]: I1124 07:18:39.539204 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d435941-9f2b-44e0-b18f-33d57d505e39" containerName="registry-server" Nov 24 07:18:39 crc kubenswrapper[4809]: E1124 07:18:39.539222 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d435941-9f2b-44e0-b18f-33d57d505e39" containerName="extract-content" Nov 24 07:18:39 crc kubenswrapper[4809]: I1124 07:18:39.539229 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d435941-9f2b-44e0-b18f-33d57d505e39" containerName="extract-content" Nov 24 07:18:39 crc kubenswrapper[4809]: E1124 07:18:39.539248 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d435941-9f2b-44e0-b18f-33d57d505e39" containerName="extract-utilities" Nov 24 07:18:39 crc kubenswrapper[4809]: I1124 07:18:39.539254 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d435941-9f2b-44e0-b18f-33d57d505e39" containerName="extract-utilities" Nov 24 07:18:39 crc kubenswrapper[4809]: I1124 07:18:39.539434 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d435941-9f2b-44e0-b18f-33d57d505e39" containerName="registry-server" Nov 24 07:18:39 crc kubenswrapper[4809]: I1124 07:18:39.540764 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k687t" Nov 24 07:18:39 crc kubenswrapper[4809]: I1124 07:18:39.554488 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k687t"] Nov 24 07:18:39 crc kubenswrapper[4809]: I1124 07:18:39.583950 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rjlcg" Nov 24 07:18:39 crc kubenswrapper[4809]: I1124 07:18:39.704995 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ce662b0-4e84-4e46-a27e-ff419a37da80-catalog-content\") pod \"certified-operators-k687t\" (UID: \"1ce662b0-4e84-4e46-a27e-ff419a37da80\") " pod="openshift-marketplace/certified-operators-k687t" Nov 24 07:18:39 crc kubenswrapper[4809]: I1124 07:18:39.705079 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zlq4\" (UniqueName: \"kubernetes.io/projected/1ce662b0-4e84-4e46-a27e-ff419a37da80-kube-api-access-6zlq4\") pod \"certified-operators-k687t\" (UID: \"1ce662b0-4e84-4e46-a27e-ff419a37da80\") " pod="openshift-marketplace/certified-operators-k687t" Nov 24 07:18:39 crc kubenswrapper[4809]: I1124 07:18:39.705278 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ce662b0-4e84-4e46-a27e-ff419a37da80-utilities\") pod \"certified-operators-k687t\" (UID: \"1ce662b0-4e84-4e46-a27e-ff419a37da80\") " pod="openshift-marketplace/certified-operators-k687t" Nov 24 07:18:39 crc kubenswrapper[4809]: I1124 07:18:39.806604 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ce662b0-4e84-4e46-a27e-ff419a37da80-utilities\") pod \"certified-operators-k687t\" (UID: \"1ce662b0-4e84-4e46-a27e-ff419a37da80\") " pod="openshift-marketplace/certified-operators-k687t" Nov 24 07:18:39 crc kubenswrapper[4809]: I1124 07:18:39.806914 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ce662b0-4e84-4e46-a27e-ff419a37da80-catalog-content\") pod \"certified-operators-k687t\" (UID: \"1ce662b0-4e84-4e46-a27e-ff419a37da80\") " pod="openshift-marketplace/certified-operators-k687t" Nov 24 07:18:39 crc kubenswrapper[4809]: I1124 07:18:39.807017 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zlq4\" (UniqueName: \"kubernetes.io/projected/1ce662b0-4e84-4e46-a27e-ff419a37da80-kube-api-access-6zlq4\") pod \"certified-operators-k687t\" (UID: \"1ce662b0-4e84-4e46-a27e-ff419a37da80\") " pod="openshift-marketplace/certified-operators-k687t" Nov 24 07:18:39 crc kubenswrapper[4809]: I1124 07:18:39.807264 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ce662b0-4e84-4e46-a27e-ff419a37da80-utilities\") pod \"certified-operators-k687t\" (UID: \"1ce662b0-4e84-4e46-a27e-ff419a37da80\") " pod="openshift-marketplace/certified-operators-k687t" Nov 24 07:18:39 crc kubenswrapper[4809]: I1124 07:18:39.807420 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ce662b0-4e84-4e46-a27e-ff419a37da80-catalog-content\") pod \"certified-operators-k687t\" (UID: \"1ce662b0-4e84-4e46-a27e-ff419a37da80\") " pod="openshift-marketplace/certified-operators-k687t" Nov 24 07:18:39 crc kubenswrapper[4809]: I1124 07:18:39.828722 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zlq4\" (UniqueName: \"kubernetes.io/projected/1ce662b0-4e84-4e46-a27e-ff419a37da80-kube-api-access-6zlq4\") pod \"certified-operators-k687t\" (UID: \"1ce662b0-4e84-4e46-a27e-ff419a37da80\") " pod="openshift-marketplace/certified-operators-k687t" Nov 24 07:18:39 crc kubenswrapper[4809]: I1124 07:18:39.871892 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k687t" Nov 24 07:18:40 crc kubenswrapper[4809]: I1124 07:18:40.408277 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k687t"] Nov 24 07:18:40 crc kubenswrapper[4809]: I1124 07:18:40.467986 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k687t" event={"ID":"1ce662b0-4e84-4e46-a27e-ff419a37da80","Type":"ContainerStarted","Data":"9f72db8542be31b256051fdd4655cb9243b0ac2e625c3826d66d1be03730625a"} Nov 24 07:18:40 crc kubenswrapper[4809]: I1124 07:18:40.536016 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rjlcg" Nov 24 07:18:41 crc kubenswrapper[4809]: I1124 07:18:41.480942 4809 generic.go:334] "Generic (PLEG): container finished" podID="1ce662b0-4e84-4e46-a27e-ff419a37da80" containerID="3a4e8b0ce8e9734f0380e3c92fc315242fc719f065f5f8482ed28282119683af" exitCode=0 Nov 24 07:18:41 crc kubenswrapper[4809]: I1124 07:18:41.481088 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k687t" event={"ID":"1ce662b0-4e84-4e46-a27e-ff419a37da80","Type":"ContainerDied","Data":"3a4e8b0ce8e9734f0380e3c92fc315242fc719f065f5f8482ed28282119683af"} Nov 24 07:18:41 crc kubenswrapper[4809]: I1124 07:18:41.935299 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rjlcg"] Nov 24 07:18:42 crc kubenswrapper[4809]: I1124 07:18:42.494174 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k687t" event={"ID":"1ce662b0-4e84-4e46-a27e-ff419a37da80","Type":"ContainerStarted","Data":"a356de740d6071f88f61439ed07338f3dbd8a192bfbe12689ffe710c15354228"} Nov 24 07:18:42 crc kubenswrapper[4809]: I1124 07:18:42.494319 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rjlcg" podUID="21b5fc68-cb78-4a1f-b924-25720d84deb4" containerName="registry-server" containerID="cri-o://526e9f59e535aa0bb976427c59b6c8ed02e433d6a478d9cbfc71b7e0941cb925" gracePeriod=2 Nov 24 07:18:42 crc kubenswrapper[4809]: I1124 07:18:42.956152 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rjlcg" Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.076603 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21b5fc68-cb78-4a1f-b924-25720d84deb4-utilities\") pod \"21b5fc68-cb78-4a1f-b924-25720d84deb4\" (UID: \"21b5fc68-cb78-4a1f-b924-25720d84deb4\") " Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.076757 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21b5fc68-cb78-4a1f-b924-25720d84deb4-catalog-content\") pod \"21b5fc68-cb78-4a1f-b924-25720d84deb4\" (UID: \"21b5fc68-cb78-4a1f-b924-25720d84deb4\") " Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.076793 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k97r\" (UniqueName: \"kubernetes.io/projected/21b5fc68-cb78-4a1f-b924-25720d84deb4-kube-api-access-2k97r\") pod \"21b5fc68-cb78-4a1f-b924-25720d84deb4\" (UID: \"21b5fc68-cb78-4a1f-b924-25720d84deb4\") " Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.077716 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21b5fc68-cb78-4a1f-b924-25720d84deb4-utilities" (OuterVolumeSpecName: "utilities") pod "21b5fc68-cb78-4a1f-b924-25720d84deb4" (UID: "21b5fc68-cb78-4a1f-b924-25720d84deb4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.087343 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21b5fc68-cb78-4a1f-b924-25720d84deb4-kube-api-access-2k97r" (OuterVolumeSpecName: "kube-api-access-2k97r") pod "21b5fc68-cb78-4a1f-b924-25720d84deb4" (UID: "21b5fc68-cb78-4a1f-b924-25720d84deb4"). InnerVolumeSpecName "kube-api-access-2k97r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.141391 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21b5fc68-cb78-4a1f-b924-25720d84deb4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "21b5fc68-cb78-4a1f-b924-25720d84deb4" (UID: "21b5fc68-cb78-4a1f-b924-25720d84deb4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.179915 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21b5fc68-cb78-4a1f-b924-25720d84deb4-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.179956 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21b5fc68-cb78-4a1f-b924-25720d84deb4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.180024 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k97r\" (UniqueName: \"kubernetes.io/projected/21b5fc68-cb78-4a1f-b924-25720d84deb4-kube-api-access-2k97r\") on node \"crc\" DevicePath \"\"" Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.507101 4809 generic.go:334] "Generic (PLEG): container finished" podID="21b5fc68-cb78-4a1f-b924-25720d84deb4" containerID="526e9f59e535aa0bb976427c59b6c8ed02e433d6a478d9cbfc71b7e0941cb925" exitCode=0 Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.507203 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjlcg" event={"ID":"21b5fc68-cb78-4a1f-b924-25720d84deb4","Type":"ContainerDied","Data":"526e9f59e535aa0bb976427c59b6c8ed02e433d6a478d9cbfc71b7e0941cb925"} Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.507254 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rjlcg" Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.507268 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjlcg" event={"ID":"21b5fc68-cb78-4a1f-b924-25720d84deb4","Type":"ContainerDied","Data":"1f72d6e6e439a59c1ae6b2b5fdddc8532eab9f7ff15572e3f8b3a94478ffaafa"} Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.507300 4809 scope.go:117] "RemoveContainer" containerID="526e9f59e535aa0bb976427c59b6c8ed02e433d6a478d9cbfc71b7e0941cb925" Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.512298 4809 generic.go:334] "Generic (PLEG): container finished" podID="1ce662b0-4e84-4e46-a27e-ff419a37da80" containerID="a356de740d6071f88f61439ed07338f3dbd8a192bfbe12689ffe710c15354228" exitCode=0 Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.512364 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k687t" event={"ID":"1ce662b0-4e84-4e46-a27e-ff419a37da80","Type":"ContainerDied","Data":"a356de740d6071f88f61439ed07338f3dbd8a192bfbe12689ffe710c15354228"} Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.545772 4809 scope.go:117] "RemoveContainer" containerID="5a797d1d3358a0dc27ad771e4795628b9884fad19c9f6508b8e4cae9a716eb62" Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.572903 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rjlcg"] Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.586502 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rjlcg"] Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.586531 4809 scope.go:117] "RemoveContainer" containerID="09d2e695d2631a38547b6b756c78ebcfb9b7a4db9c9a376ba6706803a670361c" Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.637743 4809 scope.go:117] "RemoveContainer" containerID="526e9f59e535aa0bb976427c59b6c8ed02e433d6a478d9cbfc71b7e0941cb925" Nov 24 07:18:43 crc kubenswrapper[4809]: E1124 07:18:43.638197 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"526e9f59e535aa0bb976427c59b6c8ed02e433d6a478d9cbfc71b7e0941cb925\": container with ID starting with 526e9f59e535aa0bb976427c59b6c8ed02e433d6a478d9cbfc71b7e0941cb925 not found: ID does not exist" containerID="526e9f59e535aa0bb976427c59b6c8ed02e433d6a478d9cbfc71b7e0941cb925" Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.638228 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"526e9f59e535aa0bb976427c59b6c8ed02e433d6a478d9cbfc71b7e0941cb925"} err="failed to get container status \"526e9f59e535aa0bb976427c59b6c8ed02e433d6a478d9cbfc71b7e0941cb925\": rpc error: code = NotFound desc = could not find container \"526e9f59e535aa0bb976427c59b6c8ed02e433d6a478d9cbfc71b7e0941cb925\": container with ID starting with 526e9f59e535aa0bb976427c59b6c8ed02e433d6a478d9cbfc71b7e0941cb925 not found: ID does not exist" Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.638251 4809 scope.go:117] "RemoveContainer" containerID="5a797d1d3358a0dc27ad771e4795628b9884fad19c9f6508b8e4cae9a716eb62" Nov 24 07:18:43 crc kubenswrapper[4809]: E1124 07:18:43.638745 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a797d1d3358a0dc27ad771e4795628b9884fad19c9f6508b8e4cae9a716eb62\": container with ID starting with 5a797d1d3358a0dc27ad771e4795628b9884fad19c9f6508b8e4cae9a716eb62 not found: ID does not exist" containerID="5a797d1d3358a0dc27ad771e4795628b9884fad19c9f6508b8e4cae9a716eb62" Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.638769 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a797d1d3358a0dc27ad771e4795628b9884fad19c9f6508b8e4cae9a716eb62"} err="failed to get container status \"5a797d1d3358a0dc27ad771e4795628b9884fad19c9f6508b8e4cae9a716eb62\": rpc error: code = NotFound desc = could not find container \"5a797d1d3358a0dc27ad771e4795628b9884fad19c9f6508b8e4cae9a716eb62\": container with ID starting with 5a797d1d3358a0dc27ad771e4795628b9884fad19c9f6508b8e4cae9a716eb62 not found: ID does not exist" Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.638782 4809 scope.go:117] "RemoveContainer" containerID="09d2e695d2631a38547b6b756c78ebcfb9b7a4db9c9a376ba6706803a670361c" Nov 24 07:18:43 crc kubenswrapper[4809]: E1124 07:18:43.639245 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09d2e695d2631a38547b6b756c78ebcfb9b7a4db9c9a376ba6706803a670361c\": container with ID starting with 09d2e695d2631a38547b6b756c78ebcfb9b7a4db9c9a376ba6706803a670361c not found: ID does not exist" containerID="09d2e695d2631a38547b6b756c78ebcfb9b7a4db9c9a376ba6706803a670361c" Nov 24 07:18:43 crc kubenswrapper[4809]: I1124 07:18:43.639293 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09d2e695d2631a38547b6b756c78ebcfb9b7a4db9c9a376ba6706803a670361c"} err="failed to get container status \"09d2e695d2631a38547b6b756c78ebcfb9b7a4db9c9a376ba6706803a670361c\": rpc error: code = NotFound desc = could not find container \"09d2e695d2631a38547b6b756c78ebcfb9b7a4db9c9a376ba6706803a670361c\": container with ID starting with 09d2e695d2631a38547b6b756c78ebcfb9b7a4db9c9a376ba6706803a670361c not found: ID does not exist" Nov 24 07:18:44 crc kubenswrapper[4809]: I1124 07:18:44.533607 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k687t" event={"ID":"1ce662b0-4e84-4e46-a27e-ff419a37da80","Type":"ContainerStarted","Data":"c8d39c8b5e3d309c1b107075a5c594200345a58c99835a76e989b97a3b319315"} Nov 24 07:18:44 crc kubenswrapper[4809]: I1124 07:18:44.571438 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k687t" podStartSLOduration=3.087818152 podStartE2EDuration="5.571409944s" podCreationTimestamp="2025-11-24 07:18:39 +0000 UTC" firstStartedPulling="2025-11-24 07:18:41.483000606 +0000 UTC m=+1421.383592311" lastFinishedPulling="2025-11-24 07:18:43.966592398 +0000 UTC m=+1423.867184103" observedRunningTime="2025-11-24 07:18:44.55403149 +0000 UTC m=+1424.454623205" watchObservedRunningTime="2025-11-24 07:18:44.571409944 +0000 UTC m=+1424.472001659" Nov 24 07:18:44 crc kubenswrapper[4809]: I1124 07:18:44.920152 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21b5fc68-cb78-4a1f-b924-25720d84deb4" path="/var/lib/kubelet/pods/21b5fc68-cb78-4a1f-b924-25720d84deb4/volumes" Nov 24 07:18:48 crc kubenswrapper[4809]: I1124 07:18:48.044013 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:18:48 crc kubenswrapper[4809]: I1124 07:18:48.044482 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:18:49 crc kubenswrapper[4809]: I1124 07:18:49.878208 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k687t" Nov 24 07:18:49 crc kubenswrapper[4809]: I1124 07:18:49.878293 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k687t" Nov 24 07:18:49 crc kubenswrapper[4809]: I1124 07:18:49.942001 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k687t" Nov 24 07:18:50 crc kubenswrapper[4809]: I1124 07:18:50.648307 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k687t" Nov 24 07:18:50 crc kubenswrapper[4809]: I1124 07:18:50.712039 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k687t"] Nov 24 07:18:52 crc kubenswrapper[4809]: I1124 07:18:52.618040 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-k687t" podUID="1ce662b0-4e84-4e46-a27e-ff419a37da80" containerName="registry-server" containerID="cri-o://c8d39c8b5e3d309c1b107075a5c594200345a58c99835a76e989b97a3b319315" gracePeriod=2 Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.150063 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k687t" Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.294182 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ce662b0-4e84-4e46-a27e-ff419a37da80-catalog-content\") pod \"1ce662b0-4e84-4e46-a27e-ff419a37da80\" (UID: \"1ce662b0-4e84-4e46-a27e-ff419a37da80\") " Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.294251 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ce662b0-4e84-4e46-a27e-ff419a37da80-utilities\") pod \"1ce662b0-4e84-4e46-a27e-ff419a37da80\" (UID: \"1ce662b0-4e84-4e46-a27e-ff419a37da80\") " Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.294286 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zlq4\" (UniqueName: \"kubernetes.io/projected/1ce662b0-4e84-4e46-a27e-ff419a37da80-kube-api-access-6zlq4\") pod \"1ce662b0-4e84-4e46-a27e-ff419a37da80\" (UID: \"1ce662b0-4e84-4e46-a27e-ff419a37da80\") " Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.295495 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ce662b0-4e84-4e46-a27e-ff419a37da80-utilities" (OuterVolumeSpecName: "utilities") pod "1ce662b0-4e84-4e46-a27e-ff419a37da80" (UID: "1ce662b0-4e84-4e46-a27e-ff419a37da80"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.302186 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ce662b0-4e84-4e46-a27e-ff419a37da80-kube-api-access-6zlq4" (OuterVolumeSpecName: "kube-api-access-6zlq4") pod "1ce662b0-4e84-4e46-a27e-ff419a37da80" (UID: "1ce662b0-4e84-4e46-a27e-ff419a37da80"). InnerVolumeSpecName "kube-api-access-6zlq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.333879 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ce662b0-4e84-4e46-a27e-ff419a37da80-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1ce662b0-4e84-4e46-a27e-ff419a37da80" (UID: "1ce662b0-4e84-4e46-a27e-ff419a37da80"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.397883 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ce662b0-4e84-4e46-a27e-ff419a37da80-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.397933 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ce662b0-4e84-4e46-a27e-ff419a37da80-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.397983 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zlq4\" (UniqueName: \"kubernetes.io/projected/1ce662b0-4e84-4e46-a27e-ff419a37da80-kube-api-access-6zlq4\") on node \"crc\" DevicePath \"\"" Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.637445 4809 generic.go:334] "Generic (PLEG): container finished" podID="1ce662b0-4e84-4e46-a27e-ff419a37da80" containerID="c8d39c8b5e3d309c1b107075a5c594200345a58c99835a76e989b97a3b319315" exitCode=0 Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.637510 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k687t" event={"ID":"1ce662b0-4e84-4e46-a27e-ff419a37da80","Type":"ContainerDied","Data":"c8d39c8b5e3d309c1b107075a5c594200345a58c99835a76e989b97a3b319315"} Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.637553 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k687t" event={"ID":"1ce662b0-4e84-4e46-a27e-ff419a37da80","Type":"ContainerDied","Data":"9f72db8542be31b256051fdd4655cb9243b0ac2e625c3826d66d1be03730625a"} Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.637589 4809 scope.go:117] "RemoveContainer" containerID="c8d39c8b5e3d309c1b107075a5c594200345a58c99835a76e989b97a3b319315" Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.637807 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k687t" Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.678590 4809 scope.go:117] "RemoveContainer" containerID="a356de740d6071f88f61439ed07338f3dbd8a192bfbe12689ffe710c15354228" Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.716897 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k687t"] Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.724827 4809 scope.go:117] "RemoveContainer" containerID="3a4e8b0ce8e9734f0380e3c92fc315242fc719f065f5f8482ed28282119683af" Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.725315 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-k687t"] Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.788959 4809 scope.go:117] "RemoveContainer" containerID="c8d39c8b5e3d309c1b107075a5c594200345a58c99835a76e989b97a3b319315" Nov 24 07:18:53 crc kubenswrapper[4809]: E1124 07:18:53.789802 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8d39c8b5e3d309c1b107075a5c594200345a58c99835a76e989b97a3b319315\": container with ID starting with c8d39c8b5e3d309c1b107075a5c594200345a58c99835a76e989b97a3b319315 not found: ID does not exist" containerID="c8d39c8b5e3d309c1b107075a5c594200345a58c99835a76e989b97a3b319315" Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.789879 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8d39c8b5e3d309c1b107075a5c594200345a58c99835a76e989b97a3b319315"} err="failed to get container status \"c8d39c8b5e3d309c1b107075a5c594200345a58c99835a76e989b97a3b319315\": rpc error: code = NotFound desc = could not find container \"c8d39c8b5e3d309c1b107075a5c594200345a58c99835a76e989b97a3b319315\": container with ID starting with c8d39c8b5e3d309c1b107075a5c594200345a58c99835a76e989b97a3b319315 not found: ID does not exist" Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.789918 4809 scope.go:117] "RemoveContainer" containerID="a356de740d6071f88f61439ed07338f3dbd8a192bfbe12689ffe710c15354228" Nov 24 07:18:53 crc kubenswrapper[4809]: E1124 07:18:53.790535 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a356de740d6071f88f61439ed07338f3dbd8a192bfbe12689ffe710c15354228\": container with ID starting with a356de740d6071f88f61439ed07338f3dbd8a192bfbe12689ffe710c15354228 not found: ID does not exist" containerID="a356de740d6071f88f61439ed07338f3dbd8a192bfbe12689ffe710c15354228" Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.790609 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a356de740d6071f88f61439ed07338f3dbd8a192bfbe12689ffe710c15354228"} err="failed to get container status \"a356de740d6071f88f61439ed07338f3dbd8a192bfbe12689ffe710c15354228\": rpc error: code = NotFound desc = could not find container \"a356de740d6071f88f61439ed07338f3dbd8a192bfbe12689ffe710c15354228\": container with ID starting with a356de740d6071f88f61439ed07338f3dbd8a192bfbe12689ffe710c15354228 not found: ID does not exist" Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.790660 4809 scope.go:117] "RemoveContainer" containerID="3a4e8b0ce8e9734f0380e3c92fc315242fc719f065f5f8482ed28282119683af" Nov 24 07:18:53 crc kubenswrapper[4809]: E1124 07:18:53.791242 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a4e8b0ce8e9734f0380e3c92fc315242fc719f065f5f8482ed28282119683af\": container with ID starting with 3a4e8b0ce8e9734f0380e3c92fc315242fc719f065f5f8482ed28282119683af not found: ID does not exist" containerID="3a4e8b0ce8e9734f0380e3c92fc315242fc719f065f5f8482ed28282119683af" Nov 24 07:18:53 crc kubenswrapper[4809]: I1124 07:18:53.791310 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a4e8b0ce8e9734f0380e3c92fc315242fc719f065f5f8482ed28282119683af"} err="failed to get container status \"3a4e8b0ce8e9734f0380e3c92fc315242fc719f065f5f8482ed28282119683af\": rpc error: code = NotFound desc = could not find container \"3a4e8b0ce8e9734f0380e3c92fc315242fc719f065f5f8482ed28282119683af\": container with ID starting with 3a4e8b0ce8e9734f0380e3c92fc315242fc719f065f5f8482ed28282119683af not found: ID does not exist" Nov 24 07:18:54 crc kubenswrapper[4809]: I1124 07:18:54.902439 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ce662b0-4e84-4e46-a27e-ff419a37da80" path="/var/lib/kubelet/pods/1ce662b0-4e84-4e46-a27e-ff419a37da80/volumes" Nov 24 07:19:07 crc kubenswrapper[4809]: I1124 07:19:07.687766 4809 scope.go:117] "RemoveContainer" containerID="fc7217ce9dfe93fc6257d290ff3397837947db3a1b27a521147843c08d49e3e1" Nov 24 07:19:07 crc kubenswrapper[4809]: I1124 07:19:07.721373 4809 scope.go:117] "RemoveContainer" containerID="4891063f6682c4b2811a0c696b9c08b9dafd90c63f30e8767f8a00f915cc236a" Nov 24 07:19:18 crc kubenswrapper[4809]: I1124 07:19:18.043253 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:19:18 crc kubenswrapper[4809]: I1124 07:19:18.044033 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:19:36 crc kubenswrapper[4809]: I1124 07:19:36.136185 4809 generic.go:334] "Generic (PLEG): container finished" podID="bad47712-c44f-4799-bfe2-6c95ea3b513f" containerID="f38b4bf098bbc323eaf151c0fa60f6c3b147b8d534efe7dfbc08d219cec584e1" exitCode=0 Nov 24 07:19:36 crc kubenswrapper[4809]: I1124 07:19:36.136298 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28" event={"ID":"bad47712-c44f-4799-bfe2-6c95ea3b513f","Type":"ContainerDied","Data":"f38b4bf098bbc323eaf151c0fa60f6c3b147b8d534efe7dfbc08d219cec584e1"} Nov 24 07:19:37 crc kubenswrapper[4809]: I1124 07:19:37.686625 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28" Nov 24 07:19:37 crc kubenswrapper[4809]: I1124 07:19:37.743643 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvz8w\" (UniqueName: \"kubernetes.io/projected/bad47712-c44f-4799-bfe2-6c95ea3b513f-kube-api-access-bvz8w\") pod \"bad47712-c44f-4799-bfe2-6c95ea3b513f\" (UID: \"bad47712-c44f-4799-bfe2-6c95ea3b513f\") " Nov 24 07:19:37 crc kubenswrapper[4809]: I1124 07:19:37.743718 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bad47712-c44f-4799-bfe2-6c95ea3b513f-ssh-key\") pod \"bad47712-c44f-4799-bfe2-6c95ea3b513f\" (UID: \"bad47712-c44f-4799-bfe2-6c95ea3b513f\") " Nov 24 07:19:37 crc kubenswrapper[4809]: I1124 07:19:37.743745 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad47712-c44f-4799-bfe2-6c95ea3b513f-bootstrap-combined-ca-bundle\") pod \"bad47712-c44f-4799-bfe2-6c95ea3b513f\" (UID: \"bad47712-c44f-4799-bfe2-6c95ea3b513f\") " Nov 24 07:19:37 crc kubenswrapper[4809]: I1124 07:19:37.743807 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bad47712-c44f-4799-bfe2-6c95ea3b513f-inventory\") pod \"bad47712-c44f-4799-bfe2-6c95ea3b513f\" (UID: \"bad47712-c44f-4799-bfe2-6c95ea3b513f\") " Nov 24 07:19:37 crc kubenswrapper[4809]: I1124 07:19:37.750451 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bad47712-c44f-4799-bfe2-6c95ea3b513f-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "bad47712-c44f-4799-bfe2-6c95ea3b513f" (UID: "bad47712-c44f-4799-bfe2-6c95ea3b513f"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:19:37 crc kubenswrapper[4809]: I1124 07:19:37.750634 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bad47712-c44f-4799-bfe2-6c95ea3b513f-kube-api-access-bvz8w" (OuterVolumeSpecName: "kube-api-access-bvz8w") pod "bad47712-c44f-4799-bfe2-6c95ea3b513f" (UID: "bad47712-c44f-4799-bfe2-6c95ea3b513f"). InnerVolumeSpecName "kube-api-access-bvz8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:19:37 crc kubenswrapper[4809]: I1124 07:19:37.778596 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bad47712-c44f-4799-bfe2-6c95ea3b513f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bad47712-c44f-4799-bfe2-6c95ea3b513f" (UID: "bad47712-c44f-4799-bfe2-6c95ea3b513f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:19:37 crc kubenswrapper[4809]: I1124 07:19:37.796843 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bad47712-c44f-4799-bfe2-6c95ea3b513f-inventory" (OuterVolumeSpecName: "inventory") pod "bad47712-c44f-4799-bfe2-6c95ea3b513f" (UID: "bad47712-c44f-4799-bfe2-6c95ea3b513f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:19:37 crc kubenswrapper[4809]: I1124 07:19:37.847515 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvz8w\" (UniqueName: \"kubernetes.io/projected/bad47712-c44f-4799-bfe2-6c95ea3b513f-kube-api-access-bvz8w\") on node \"crc\" DevicePath \"\"" Nov 24 07:19:37 crc kubenswrapper[4809]: I1124 07:19:37.847575 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bad47712-c44f-4799-bfe2-6c95ea3b513f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 07:19:37 crc kubenswrapper[4809]: I1124 07:19:37.847598 4809 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad47712-c44f-4799-bfe2-6c95ea3b513f-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:19:37 crc kubenswrapper[4809]: I1124 07:19:37.847618 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bad47712-c44f-4799-bfe2-6c95ea3b513f-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.160352 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28" event={"ID":"bad47712-c44f-4799-bfe2-6c95ea3b513f","Type":"ContainerDied","Data":"74163b98736e7ab8efa930e3174349e09348b832cbca5dfa14f91f76ff1925d3"} Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.160393 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74163b98736e7ab8efa930e3174349e09348b832cbca5dfa14f91f76ff1925d3" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.160405 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.236685 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9"] Nov 24 07:19:38 crc kubenswrapper[4809]: E1124 07:19:38.237122 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad47712-c44f-4799-bfe2-6c95ea3b513f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.237142 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad47712-c44f-4799-bfe2-6c95ea3b513f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 07:19:38 crc kubenswrapper[4809]: E1124 07:19:38.237153 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ce662b0-4e84-4e46-a27e-ff419a37da80" containerName="registry-server" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.237159 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ce662b0-4e84-4e46-a27e-ff419a37da80" containerName="registry-server" Nov 24 07:19:38 crc kubenswrapper[4809]: E1124 07:19:38.237182 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21b5fc68-cb78-4a1f-b924-25720d84deb4" containerName="extract-content" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.237188 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="21b5fc68-cb78-4a1f-b924-25720d84deb4" containerName="extract-content" Nov 24 07:19:38 crc kubenswrapper[4809]: E1124 07:19:38.237202 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ce662b0-4e84-4e46-a27e-ff419a37da80" containerName="extract-content" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.237207 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ce662b0-4e84-4e46-a27e-ff419a37da80" containerName="extract-content" Nov 24 07:19:38 crc kubenswrapper[4809]: E1124 07:19:38.237220 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ce662b0-4e84-4e46-a27e-ff419a37da80" containerName="extract-utilities" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.237225 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ce662b0-4e84-4e46-a27e-ff419a37da80" containerName="extract-utilities" Nov 24 07:19:38 crc kubenswrapper[4809]: E1124 07:19:38.237236 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21b5fc68-cb78-4a1f-b924-25720d84deb4" containerName="registry-server" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.237241 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="21b5fc68-cb78-4a1f-b924-25720d84deb4" containerName="registry-server" Nov 24 07:19:38 crc kubenswrapper[4809]: E1124 07:19:38.237258 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21b5fc68-cb78-4a1f-b924-25720d84deb4" containerName="extract-utilities" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.237263 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="21b5fc68-cb78-4a1f-b924-25720d84deb4" containerName="extract-utilities" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.237420 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad47712-c44f-4799-bfe2-6c95ea3b513f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.237436 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ce662b0-4e84-4e46-a27e-ff419a37da80" containerName="registry-server" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.237457 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="21b5fc68-cb78-4a1f-b924-25720d84deb4" containerName="registry-server" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.238072 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.240454 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.240786 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.241124 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lx7ll" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.241375 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.254062 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9"] Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.254147 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a022199a-9d09-421e-b9f3-67d00cab4f5f-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9\" (UID: \"a022199a-9d09-421e-b9f3-67d00cab4f5f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.254181 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6zmx\" (UniqueName: \"kubernetes.io/projected/a022199a-9d09-421e-b9f3-67d00cab4f5f-kube-api-access-p6zmx\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9\" (UID: \"a022199a-9d09-421e-b9f3-67d00cab4f5f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.254238 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a022199a-9d09-421e-b9f3-67d00cab4f5f-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9\" (UID: \"a022199a-9d09-421e-b9f3-67d00cab4f5f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.355461 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a022199a-9d09-421e-b9f3-67d00cab4f5f-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9\" (UID: \"a022199a-9d09-421e-b9f3-67d00cab4f5f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.355639 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a022199a-9d09-421e-b9f3-67d00cab4f5f-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9\" (UID: \"a022199a-9d09-421e-b9f3-67d00cab4f5f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.355671 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6zmx\" (UniqueName: \"kubernetes.io/projected/a022199a-9d09-421e-b9f3-67d00cab4f5f-kube-api-access-p6zmx\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9\" (UID: \"a022199a-9d09-421e-b9f3-67d00cab4f5f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.359663 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a022199a-9d09-421e-b9f3-67d00cab4f5f-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9\" (UID: \"a022199a-9d09-421e-b9f3-67d00cab4f5f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.362539 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a022199a-9d09-421e-b9f3-67d00cab4f5f-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9\" (UID: \"a022199a-9d09-421e-b9f3-67d00cab4f5f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.376699 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6zmx\" (UniqueName: \"kubernetes.io/projected/a022199a-9d09-421e-b9f3-67d00cab4f5f-kube-api-access-p6zmx\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9\" (UID: \"a022199a-9d09-421e-b9f3-67d00cab4f5f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9" Nov 24 07:19:38 crc kubenswrapper[4809]: I1124 07:19:38.571384 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9" Nov 24 07:19:39 crc kubenswrapper[4809]: I1124 07:19:39.059398 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9"] Nov 24 07:19:39 crc kubenswrapper[4809]: I1124 07:19:39.171870 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9" event={"ID":"a022199a-9d09-421e-b9f3-67d00cab4f5f","Type":"ContainerStarted","Data":"e1c1b6651b899448f4c824466cb2342fd6b73937d9304a74aabc12412bc72ada"} Nov 24 07:19:40 crc kubenswrapper[4809]: I1124 07:19:40.183102 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9" event={"ID":"a022199a-9d09-421e-b9f3-67d00cab4f5f","Type":"ContainerStarted","Data":"d0a1e78855747314a6aed5af30a37536647c7e079108999a908c941e6b4c249f"} Nov 24 07:19:48 crc kubenswrapper[4809]: I1124 07:19:48.043314 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:19:48 crc kubenswrapper[4809]: I1124 07:19:48.044686 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:19:48 crc kubenswrapper[4809]: I1124 07:19:48.044747 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 07:19:48 crc kubenswrapper[4809]: I1124 07:19:48.045551 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e"} pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:19:48 crc kubenswrapper[4809]: I1124 07:19:48.045623 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" containerID="cri-o://e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" gracePeriod=600 Nov 24 07:19:48 crc kubenswrapper[4809]: E1124 07:19:48.171579 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:19:48 crc kubenswrapper[4809]: I1124 07:19:48.262163 4809 generic.go:334] "Generic (PLEG): container finished" podID="be8050d5-1aea-453e-a06e-87702b0e856b" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" exitCode=0 Nov 24 07:19:48 crc kubenswrapper[4809]: I1124 07:19:48.262204 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerDied","Data":"e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e"} Nov 24 07:19:48 crc kubenswrapper[4809]: I1124 07:19:48.262240 4809 scope.go:117] "RemoveContainer" containerID="11616e3dc08f43f601be45c4cf1cbd156c734d2c4a8bf401dd3c381f1cf4aefb" Nov 24 07:19:48 crc kubenswrapper[4809]: I1124 07:19:48.263176 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:19:48 crc kubenswrapper[4809]: E1124 07:19:48.263620 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:19:48 crc kubenswrapper[4809]: I1124 07:19:48.282584 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9" podStartSLOduration=9.866574698 podStartE2EDuration="10.282565733s" podCreationTimestamp="2025-11-24 07:19:38 +0000 UTC" firstStartedPulling="2025-11-24 07:19:39.062114442 +0000 UTC m=+1478.962706147" lastFinishedPulling="2025-11-24 07:19:39.478105477 +0000 UTC m=+1479.378697182" observedRunningTime="2025-11-24 07:19:40.201880399 +0000 UTC m=+1480.102472144" watchObservedRunningTime="2025-11-24 07:19:48.282565733 +0000 UTC m=+1488.183157438" Nov 24 07:19:59 crc kubenswrapper[4809]: I1124 07:19:59.892519 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:19:59 crc kubenswrapper[4809]: E1124 07:19:59.893656 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:20:07 crc kubenswrapper[4809]: I1124 07:20:07.861759 4809 scope.go:117] "RemoveContainer" containerID="4cd3d9f60485af0ee52d3e65e02fda14955ef8130cccd9b4978f7329249cf359" Nov 24 07:20:14 crc kubenswrapper[4809]: I1124 07:20:14.892366 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:20:14 crc kubenswrapper[4809]: E1124 07:20:14.893722 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:20:25 crc kubenswrapper[4809]: I1124 07:20:25.891374 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:20:25 crc kubenswrapper[4809]: E1124 07:20:25.892233 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:20:38 crc kubenswrapper[4809]: I1124 07:20:38.890913 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:20:38 crc kubenswrapper[4809]: E1124 07:20:38.892573 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:20:53 crc kubenswrapper[4809]: I1124 07:20:53.893766 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:20:53 crc kubenswrapper[4809]: E1124 07:20:53.896088 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:20:56 crc kubenswrapper[4809]: I1124 07:20:56.040218 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-zhvp2"] Nov 24 07:20:56 crc kubenswrapper[4809]: I1124 07:20:56.051256 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-6976-account-create-cl45b"] Nov 24 07:20:56 crc kubenswrapper[4809]: I1124 07:20:56.061173 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-zhvp2"] Nov 24 07:20:56 crc kubenswrapper[4809]: I1124 07:20:56.071371 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-6976-account-create-cl45b"] Nov 24 07:20:56 crc kubenswrapper[4809]: I1124 07:20:56.901671 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2adaac7b-6a74-4d88-a1f6-91c0f76c41e1" path="/var/lib/kubelet/pods/2adaac7b-6a74-4d88-a1f6-91c0f76c41e1/volumes" Nov 24 07:20:56 crc kubenswrapper[4809]: I1124 07:20:56.902816 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3f63bec-2b7d-475f-a402-0f2a2e93a7f4" path="/var/lib/kubelet/pods/a3f63bec-2b7d-475f-a402-0f2a2e93a7f4/volumes" Nov 24 07:20:59 crc kubenswrapper[4809]: I1124 07:20:59.040062 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-6b09-account-create-rwt7n"] Nov 24 07:20:59 crc kubenswrapper[4809]: I1124 07:20:59.058308 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-h29zw"] Nov 24 07:20:59 crc kubenswrapper[4809]: I1124 07:20:59.058376 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-xmx6m"] Nov 24 07:20:59 crc kubenswrapper[4809]: I1124 07:20:59.065277 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bc18-account-create-nqkws"] Nov 24 07:20:59 crc kubenswrapper[4809]: I1124 07:20:59.072951 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-xmx6m"] Nov 24 07:20:59 crc kubenswrapper[4809]: I1124 07:20:59.080319 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-h29zw"] Nov 24 07:20:59 crc kubenswrapper[4809]: I1124 07:20:59.086882 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-6b09-account-create-rwt7n"] Nov 24 07:20:59 crc kubenswrapper[4809]: I1124 07:20:59.093296 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bc18-account-create-nqkws"] Nov 24 07:21:00 crc kubenswrapper[4809]: I1124 07:21:00.910115 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27efd636-9d30-489d-aca8-f4d02581e2a8" path="/var/lib/kubelet/pods/27efd636-9d30-489d-aca8-f4d02581e2a8/volumes" Nov 24 07:21:00 crc kubenswrapper[4809]: I1124 07:21:00.911882 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68b4f942-93e8-4823-91c8-761cd177776e" path="/var/lib/kubelet/pods/68b4f942-93e8-4823-91c8-761cd177776e/volumes" Nov 24 07:21:00 crc kubenswrapper[4809]: I1124 07:21:00.913196 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b" path="/var/lib/kubelet/pods/ef2c0d94-2e0d-4bcf-a336-f0c20fb81f2b/volumes" Nov 24 07:21:00 crc kubenswrapper[4809]: I1124 07:21:00.914343 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4a7a40c-c36f-4d21-bdfa-6035613e6719" path="/var/lib/kubelet/pods/f4a7a40c-c36f-4d21-bdfa-6035613e6719/volumes" Nov 24 07:21:04 crc kubenswrapper[4809]: I1124 07:21:04.891925 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:21:04 crc kubenswrapper[4809]: E1124 07:21:04.892997 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:21:07 crc kubenswrapper[4809]: I1124 07:21:07.939038 4809 scope.go:117] "RemoveContainer" containerID="a166853a205f39cf00e12118166298d1e418274b26182a64021afd85f5c39708" Nov 24 07:21:07 crc kubenswrapper[4809]: I1124 07:21:07.991766 4809 scope.go:117] "RemoveContainer" containerID="547ecbbb34ecb4ef0dc9e6286529b13444930dd157383cbf643fa05d323fc7c4" Nov 24 07:21:08 crc kubenswrapper[4809]: I1124 07:21:08.033268 4809 scope.go:117] "RemoveContainer" containerID="9f6784d4d3e760400a2489904d937493c6e8a79e69dad97e61a457810819971b" Nov 24 07:21:08 crc kubenswrapper[4809]: I1124 07:21:08.081292 4809 scope.go:117] "RemoveContainer" containerID="190478af5ae637ae6c3799fd356d7d7ede679bb30cde7468370245ea96affb38" Nov 24 07:21:08 crc kubenswrapper[4809]: I1124 07:21:08.123324 4809 scope.go:117] "RemoveContainer" containerID="de93cbe0763eed2bf3a1bc12941fb0f8bc157e41e43e377f860e7e4554eaedc7" Nov 24 07:21:08 crc kubenswrapper[4809]: I1124 07:21:08.169695 4809 scope.go:117] "RemoveContainer" containerID="a88e6255bc3813c009d8179e27ccda73e8cf0e4f9c957b49eb72735456cb2b5a" Nov 24 07:21:16 crc kubenswrapper[4809]: I1124 07:21:16.191832 4809 generic.go:334] "Generic (PLEG): container finished" podID="a022199a-9d09-421e-b9f3-67d00cab4f5f" containerID="d0a1e78855747314a6aed5af30a37536647c7e079108999a908c941e6b4c249f" exitCode=0 Nov 24 07:21:16 crc kubenswrapper[4809]: I1124 07:21:16.191875 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9" event={"ID":"a022199a-9d09-421e-b9f3-67d00cab4f5f","Type":"ContainerDied","Data":"d0a1e78855747314a6aed5af30a37536647c7e079108999a908c941e6b4c249f"} Nov 24 07:21:17 crc kubenswrapper[4809]: I1124 07:21:17.626433 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9" Nov 24 07:21:17 crc kubenswrapper[4809]: I1124 07:21:17.706441 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a022199a-9d09-421e-b9f3-67d00cab4f5f-inventory\") pod \"a022199a-9d09-421e-b9f3-67d00cab4f5f\" (UID: \"a022199a-9d09-421e-b9f3-67d00cab4f5f\") " Nov 24 07:21:17 crc kubenswrapper[4809]: I1124 07:21:17.706490 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a022199a-9d09-421e-b9f3-67d00cab4f5f-ssh-key\") pod \"a022199a-9d09-421e-b9f3-67d00cab4f5f\" (UID: \"a022199a-9d09-421e-b9f3-67d00cab4f5f\") " Nov 24 07:21:17 crc kubenswrapper[4809]: I1124 07:21:17.706539 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6zmx\" (UniqueName: \"kubernetes.io/projected/a022199a-9d09-421e-b9f3-67d00cab4f5f-kube-api-access-p6zmx\") pod \"a022199a-9d09-421e-b9f3-67d00cab4f5f\" (UID: \"a022199a-9d09-421e-b9f3-67d00cab4f5f\") " Nov 24 07:21:17 crc kubenswrapper[4809]: I1124 07:21:17.731262 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a022199a-9d09-421e-b9f3-67d00cab4f5f-kube-api-access-p6zmx" (OuterVolumeSpecName: "kube-api-access-p6zmx") pod "a022199a-9d09-421e-b9f3-67d00cab4f5f" (UID: "a022199a-9d09-421e-b9f3-67d00cab4f5f"). InnerVolumeSpecName "kube-api-access-p6zmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:21:17 crc kubenswrapper[4809]: I1124 07:21:17.763152 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a022199a-9d09-421e-b9f3-67d00cab4f5f-inventory" (OuterVolumeSpecName: "inventory") pod "a022199a-9d09-421e-b9f3-67d00cab4f5f" (UID: "a022199a-9d09-421e-b9f3-67d00cab4f5f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:21:17 crc kubenswrapper[4809]: I1124 07:21:17.767727 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a022199a-9d09-421e-b9f3-67d00cab4f5f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a022199a-9d09-421e-b9f3-67d00cab4f5f" (UID: "a022199a-9d09-421e-b9f3-67d00cab4f5f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:21:17 crc kubenswrapper[4809]: I1124 07:21:17.809291 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a022199a-9d09-421e-b9f3-67d00cab4f5f-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 07:21:17 crc kubenswrapper[4809]: I1124 07:21:17.809321 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a022199a-9d09-421e-b9f3-67d00cab4f5f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 07:21:17 crc kubenswrapper[4809]: I1124 07:21:17.809330 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6zmx\" (UniqueName: \"kubernetes.io/projected/a022199a-9d09-421e-b9f3-67d00cab4f5f-kube-api-access-p6zmx\") on node \"crc\" DevicePath \"\"" Nov 24 07:21:18 crc kubenswrapper[4809]: I1124 07:21:18.224673 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9" event={"ID":"a022199a-9d09-421e-b9f3-67d00cab4f5f","Type":"ContainerDied","Data":"e1c1b6651b899448f4c824466cb2342fd6b73937d9304a74aabc12412bc72ada"} Nov 24 07:21:18 crc kubenswrapper[4809]: I1124 07:21:18.224751 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1c1b6651b899448f4c824466cb2342fd6b73937d9304a74aabc12412bc72ada" Nov 24 07:21:18 crc kubenswrapper[4809]: I1124 07:21:18.224831 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9" Nov 24 07:21:18 crc kubenswrapper[4809]: I1124 07:21:18.290517 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k"] Nov 24 07:21:18 crc kubenswrapper[4809]: E1124 07:21:18.291019 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a022199a-9d09-421e-b9f3-67d00cab4f5f" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 24 07:21:18 crc kubenswrapper[4809]: I1124 07:21:18.291040 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a022199a-9d09-421e-b9f3-67d00cab4f5f" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 24 07:21:18 crc kubenswrapper[4809]: I1124 07:21:18.291271 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a022199a-9d09-421e-b9f3-67d00cab4f5f" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 24 07:21:18 crc kubenswrapper[4809]: I1124 07:21:18.293444 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k" Nov 24 07:21:18 crc kubenswrapper[4809]: I1124 07:21:18.296955 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 07:21:18 crc kubenswrapper[4809]: I1124 07:21:18.297165 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lx7ll" Nov 24 07:21:18 crc kubenswrapper[4809]: I1124 07:21:18.297489 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 07:21:18 crc kubenswrapper[4809]: I1124 07:21:18.297768 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 07:21:18 crc kubenswrapper[4809]: I1124 07:21:18.302257 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k"] Nov 24 07:21:18 crc kubenswrapper[4809]: I1124 07:21:18.420314 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nq4l\" (UniqueName: \"kubernetes.io/projected/8a942ac4-7652-45aa-b238-3aff50848195-kube-api-access-9nq4l\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k\" (UID: \"8a942ac4-7652-45aa-b238-3aff50848195\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k" Nov 24 07:21:18 crc kubenswrapper[4809]: I1124 07:21:18.420457 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a942ac4-7652-45aa-b238-3aff50848195-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k\" (UID: \"8a942ac4-7652-45aa-b238-3aff50848195\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k" Nov 24 07:21:18 crc kubenswrapper[4809]: I1124 07:21:18.420572 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a942ac4-7652-45aa-b238-3aff50848195-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k\" (UID: \"8a942ac4-7652-45aa-b238-3aff50848195\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k" Nov 24 07:21:18 crc kubenswrapper[4809]: I1124 07:21:18.522311 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nq4l\" (UniqueName: \"kubernetes.io/projected/8a942ac4-7652-45aa-b238-3aff50848195-kube-api-access-9nq4l\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k\" (UID: \"8a942ac4-7652-45aa-b238-3aff50848195\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k" Nov 24 07:21:18 crc kubenswrapper[4809]: I1124 07:21:18.522607 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a942ac4-7652-45aa-b238-3aff50848195-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k\" (UID: \"8a942ac4-7652-45aa-b238-3aff50848195\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k" Nov 24 07:21:18 crc kubenswrapper[4809]: I1124 07:21:18.522726 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a942ac4-7652-45aa-b238-3aff50848195-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k\" (UID: \"8a942ac4-7652-45aa-b238-3aff50848195\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k" Nov 24 07:21:18 crc kubenswrapper[4809]: I1124 07:21:18.525920 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a942ac4-7652-45aa-b238-3aff50848195-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k\" (UID: \"8a942ac4-7652-45aa-b238-3aff50848195\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k" Nov 24 07:21:18 crc kubenswrapper[4809]: I1124 07:21:18.527464 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a942ac4-7652-45aa-b238-3aff50848195-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k\" (UID: \"8a942ac4-7652-45aa-b238-3aff50848195\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k" Nov 24 07:21:18 crc kubenswrapper[4809]: I1124 07:21:18.539002 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nq4l\" (UniqueName: \"kubernetes.io/projected/8a942ac4-7652-45aa-b238-3aff50848195-kube-api-access-9nq4l\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k\" (UID: \"8a942ac4-7652-45aa-b238-3aff50848195\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k" Nov 24 07:21:18 crc kubenswrapper[4809]: I1124 07:21:18.654135 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k" Nov 24 07:21:18 crc kubenswrapper[4809]: I1124 07:21:18.891167 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:21:18 crc kubenswrapper[4809]: E1124 07:21:18.891753 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:21:19 crc kubenswrapper[4809]: I1124 07:21:19.184482 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k"] Nov 24 07:21:19 crc kubenswrapper[4809]: I1124 07:21:19.193183 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 07:21:19 crc kubenswrapper[4809]: I1124 07:21:19.236347 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k" event={"ID":"8a942ac4-7652-45aa-b238-3aff50848195","Type":"ContainerStarted","Data":"4181142860d70def7bcb809f8a3ff53926958251d4faac3bfc775af86fbdb19a"} Nov 24 07:21:20 crc kubenswrapper[4809]: I1124 07:21:20.251447 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k" event={"ID":"8a942ac4-7652-45aa-b238-3aff50848195","Type":"ContainerStarted","Data":"46cdff963788b6b4d1adf75faafac83d22441bc48596bd110df7dfddd637e7fc"} Nov 24 07:21:20 crc kubenswrapper[4809]: I1124 07:21:20.287741 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k" podStartSLOduration=1.8808530289999998 podStartE2EDuration="2.287717229s" podCreationTimestamp="2025-11-24 07:21:18 +0000 UTC" firstStartedPulling="2025-11-24 07:21:19.19292026 +0000 UTC m=+1579.093511965" lastFinishedPulling="2025-11-24 07:21:19.59978445 +0000 UTC m=+1579.500376165" observedRunningTime="2025-11-24 07:21:20.274489116 +0000 UTC m=+1580.175080841" watchObservedRunningTime="2025-11-24 07:21:20.287717229 +0000 UTC m=+1580.188308934" Nov 24 07:21:22 crc kubenswrapper[4809]: I1124 07:21:22.038048 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-lfw7d"] Nov 24 07:21:22 crc kubenswrapper[4809]: I1124 07:21:22.046273 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-lfw7d"] Nov 24 07:21:22 crc kubenswrapper[4809]: I1124 07:21:22.906306 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="952df16f-264d-49ca-9281-9f3b95f8273d" path="/var/lib/kubelet/pods/952df16f-264d-49ca-9281-9f3b95f8273d/volumes" Nov 24 07:21:29 crc kubenswrapper[4809]: I1124 07:21:29.891702 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:21:29 crc kubenswrapper[4809]: E1124 07:21:29.892621 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:21:30 crc kubenswrapper[4809]: I1124 07:21:30.036239 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-vxq56"] Nov 24 07:21:30 crc kubenswrapper[4809]: I1124 07:21:30.045802 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-00f1-account-create-pfn7s"] Nov 24 07:21:30 crc kubenswrapper[4809]: I1124 07:21:30.058142 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-pz5bj"] Nov 24 07:21:30 crc kubenswrapper[4809]: I1124 07:21:30.068516 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-00f1-account-create-pfn7s"] Nov 24 07:21:30 crc kubenswrapper[4809]: I1124 07:21:30.075821 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-e4c1-account-create-4f5hg"] Nov 24 07:21:30 crc kubenswrapper[4809]: I1124 07:21:30.082099 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-e4c1-account-create-4f5hg"] Nov 24 07:21:30 crc kubenswrapper[4809]: I1124 07:21:30.089277 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-pz5bj"] Nov 24 07:21:30 crc kubenswrapper[4809]: I1124 07:21:30.097625 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-vxq56"] Nov 24 07:21:30 crc kubenswrapper[4809]: I1124 07:21:30.105342 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-dnl8c"] Nov 24 07:21:30 crc kubenswrapper[4809]: I1124 07:21:30.112203 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-2540-account-create-w7dp4"] Nov 24 07:21:30 crc kubenswrapper[4809]: I1124 07:21:30.118290 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-2540-account-create-w7dp4"] Nov 24 07:21:30 crc kubenswrapper[4809]: I1124 07:21:30.124512 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-dnl8c"] Nov 24 07:21:30 crc kubenswrapper[4809]: I1124 07:21:30.902804 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06f78e27-776f-4506-b3f2-a348318cbd13" path="/var/lib/kubelet/pods/06f78e27-776f-4506-b3f2-a348318cbd13/volumes" Nov 24 07:21:30 crc kubenswrapper[4809]: I1124 07:21:30.903820 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bad348b-3d85-4a22-9d03-734715eedffa" path="/var/lib/kubelet/pods/1bad348b-3d85-4a22-9d03-734715eedffa/volumes" Nov 24 07:21:30 crc kubenswrapper[4809]: I1124 07:21:30.904771 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bcc82f1-7d2a-48b5-83cb-157f3d4c320c" path="/var/lib/kubelet/pods/7bcc82f1-7d2a-48b5-83cb-157f3d4c320c/volumes" Nov 24 07:21:30 crc kubenswrapper[4809]: I1124 07:21:30.905737 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c34951d-7bde-4c8e-b369-fc6505c3a752" path="/var/lib/kubelet/pods/7c34951d-7bde-4c8e-b369-fc6505c3a752/volumes" Nov 24 07:21:30 crc kubenswrapper[4809]: I1124 07:21:30.907440 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91db574f-2e79-4e56-a408-64373360aa4a" path="/var/lib/kubelet/pods/91db574f-2e79-4e56-a408-64373360aa4a/volumes" Nov 24 07:21:30 crc kubenswrapper[4809]: I1124 07:21:30.908659 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4e0697a-3b6b-4d9c-a863-9ef8a696920a" path="/var/lib/kubelet/pods/f4e0697a-3b6b-4d9c-a863-9ef8a696920a/volumes" Nov 24 07:21:35 crc kubenswrapper[4809]: I1124 07:21:35.036207 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-tbhvc"] Nov 24 07:21:35 crc kubenswrapper[4809]: I1124 07:21:35.046147 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-tbhvc"] Nov 24 07:21:36 crc kubenswrapper[4809]: I1124 07:21:36.905764 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99c73ccf-06f4-4dfb-ab11-a1b93ab653ea" path="/var/lib/kubelet/pods/99c73ccf-06f4-4dfb-ab11-a1b93ab653ea/volumes" Nov 24 07:21:42 crc kubenswrapper[4809]: I1124 07:21:42.893073 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:21:42 crc kubenswrapper[4809]: E1124 07:21:42.895666 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:21:55 crc kubenswrapper[4809]: I1124 07:21:55.891326 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:21:55 crc kubenswrapper[4809]: E1124 07:21:55.892202 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:22:07 crc kubenswrapper[4809]: I1124 07:22:07.059176 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-4l5jw"] Nov 24 07:22:07 crc kubenswrapper[4809]: I1124 07:22:07.066517 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-4l5jw"] Nov 24 07:22:07 crc kubenswrapper[4809]: I1124 07:22:07.892228 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:22:07 crc kubenswrapper[4809]: E1124 07:22:07.893084 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:22:08 crc kubenswrapper[4809]: I1124 07:22:08.367218 4809 scope.go:117] "RemoveContainer" containerID="7585ffb5da77f70ac97e781e5b676cee5a9d97f3023594a6cd2d14a0f2bf4bac" Nov 24 07:22:08 crc kubenswrapper[4809]: I1124 07:22:08.412096 4809 scope.go:117] "RemoveContainer" containerID="56318ea24cf9f8d2440b2c2e6c2eba928409aeac48663b72a34e11fd39e3dff6" Nov 24 07:22:08 crc kubenswrapper[4809]: I1124 07:22:08.446244 4809 scope.go:117] "RemoveContainer" containerID="98853418d288a056113d9c613919ff0f6958064a0058aeaff0e497d4a507db03" Nov 24 07:22:08 crc kubenswrapper[4809]: I1124 07:22:08.493750 4809 scope.go:117] "RemoveContainer" containerID="d974c3beaef64ba1537b2ec4e909be97aded61fe58f8b13cb2d636c170100f74" Nov 24 07:22:08 crc kubenswrapper[4809]: I1124 07:22:08.517034 4809 scope.go:117] "RemoveContainer" containerID="3fbf20b7ca69e9111cab29ca1b274c354c8ab7644e8fb693ff3bc3e0d58658cd" Nov 24 07:22:08 crc kubenswrapper[4809]: I1124 07:22:08.564117 4809 scope.go:117] "RemoveContainer" containerID="6e49c55d9b850ae38097aee946cdd838022b3d5f6af1325c50a9b6038bb71579" Nov 24 07:22:08 crc kubenswrapper[4809]: I1124 07:22:08.599044 4809 scope.go:117] "RemoveContainer" containerID="9e6f28cb3c7b8bc7f9374a06b3bc896f2892963f6b563dd75c5d836ebfb60db8" Nov 24 07:22:08 crc kubenswrapper[4809]: I1124 07:22:08.636594 4809 scope.go:117] "RemoveContainer" containerID="05d1220c6bdb4756a0bceb61608c2122633b928859c1d4247efbe42ba4f4d90a" Nov 24 07:22:08 crc kubenswrapper[4809]: I1124 07:22:08.904277 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cc0b19f-d90f-4460-909e-6eb644cb2955" path="/var/lib/kubelet/pods/8cc0b19f-d90f-4460-909e-6eb644cb2955/volumes" Nov 24 07:22:19 crc kubenswrapper[4809]: I1124 07:22:19.037602 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-ncjd6"] Nov 24 07:22:19 crc kubenswrapper[4809]: I1124 07:22:19.049901 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-qsvl5"] Nov 24 07:22:19 crc kubenswrapper[4809]: I1124 07:22:19.062445 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-ncjd6"] Nov 24 07:22:19 crc kubenswrapper[4809]: I1124 07:22:19.071614 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-qsvl5"] Nov 24 07:22:20 crc kubenswrapper[4809]: I1124 07:22:20.905906 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293" path="/var/lib/kubelet/pods/6a2a1588-cd17-4e48-bb3b-e6a4ef2cc293/volumes" Nov 24 07:22:20 crc kubenswrapper[4809]: I1124 07:22:20.907315 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5d93390-ea3b-43d3-83ba-4d85fb8544d4" path="/var/lib/kubelet/pods/f5d93390-ea3b-43d3-83ba-4d85fb8544d4/volumes" Nov 24 07:22:21 crc kubenswrapper[4809]: I1124 07:22:21.891514 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:22:21 crc kubenswrapper[4809]: E1124 07:22:21.891725 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:22:27 crc kubenswrapper[4809]: I1124 07:22:27.039406 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-lkj6z"] Nov 24 07:22:27 crc kubenswrapper[4809]: I1124 07:22:27.047617 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-lkj6z"] Nov 24 07:22:27 crc kubenswrapper[4809]: I1124 07:22:27.057618 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-dtg82"] Nov 24 07:22:27 crc kubenswrapper[4809]: I1124 07:22:27.066026 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-dtg82"] Nov 24 07:22:28 crc kubenswrapper[4809]: I1124 07:22:28.906847 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35dbbf41-295e-4ec9-bd02-12911c093ce8" path="/var/lib/kubelet/pods/35dbbf41-295e-4ec9-bd02-12911c093ce8/volumes" Nov 24 07:22:28 crc kubenswrapper[4809]: I1124 07:22:28.907911 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="972cb0ce-ea08-408c-a10d-ba1d14d42568" path="/var/lib/kubelet/pods/972cb0ce-ea08-408c-a10d-ba1d14d42568/volumes" Nov 24 07:22:33 crc kubenswrapper[4809]: I1124 07:22:33.891247 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:22:33 crc kubenswrapper[4809]: E1124 07:22:33.892605 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:22:34 crc kubenswrapper[4809]: I1124 07:22:34.009054 4809 generic.go:334] "Generic (PLEG): container finished" podID="8a942ac4-7652-45aa-b238-3aff50848195" containerID="46cdff963788b6b4d1adf75faafac83d22441bc48596bd110df7dfddd637e7fc" exitCode=0 Nov 24 07:22:34 crc kubenswrapper[4809]: I1124 07:22:34.009099 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k" event={"ID":"8a942ac4-7652-45aa-b238-3aff50848195","Type":"ContainerDied","Data":"46cdff963788b6b4d1adf75faafac83d22441bc48596bd110df7dfddd637e7fc"} Nov 24 07:22:35 crc kubenswrapper[4809]: I1124 07:22:35.463472 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k" Nov 24 07:22:35 crc kubenswrapper[4809]: I1124 07:22:35.590942 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a942ac4-7652-45aa-b238-3aff50848195-inventory\") pod \"8a942ac4-7652-45aa-b238-3aff50848195\" (UID: \"8a942ac4-7652-45aa-b238-3aff50848195\") " Nov 24 07:22:35 crc kubenswrapper[4809]: I1124 07:22:35.591208 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a942ac4-7652-45aa-b238-3aff50848195-ssh-key\") pod \"8a942ac4-7652-45aa-b238-3aff50848195\" (UID: \"8a942ac4-7652-45aa-b238-3aff50848195\") " Nov 24 07:22:35 crc kubenswrapper[4809]: I1124 07:22:35.591353 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nq4l\" (UniqueName: \"kubernetes.io/projected/8a942ac4-7652-45aa-b238-3aff50848195-kube-api-access-9nq4l\") pod \"8a942ac4-7652-45aa-b238-3aff50848195\" (UID: \"8a942ac4-7652-45aa-b238-3aff50848195\") " Nov 24 07:22:35 crc kubenswrapper[4809]: I1124 07:22:35.596263 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a942ac4-7652-45aa-b238-3aff50848195-kube-api-access-9nq4l" (OuterVolumeSpecName: "kube-api-access-9nq4l") pod "8a942ac4-7652-45aa-b238-3aff50848195" (UID: "8a942ac4-7652-45aa-b238-3aff50848195"). InnerVolumeSpecName "kube-api-access-9nq4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:22:35 crc kubenswrapper[4809]: I1124 07:22:35.616754 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a942ac4-7652-45aa-b238-3aff50848195-inventory" (OuterVolumeSpecName: "inventory") pod "8a942ac4-7652-45aa-b238-3aff50848195" (UID: "8a942ac4-7652-45aa-b238-3aff50848195"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:22:35 crc kubenswrapper[4809]: I1124 07:22:35.617413 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a942ac4-7652-45aa-b238-3aff50848195-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8a942ac4-7652-45aa-b238-3aff50848195" (UID: "8a942ac4-7652-45aa-b238-3aff50848195"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:22:35 crc kubenswrapper[4809]: I1124 07:22:35.693843 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a942ac4-7652-45aa-b238-3aff50848195-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 07:22:35 crc kubenswrapper[4809]: I1124 07:22:35.693886 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a942ac4-7652-45aa-b238-3aff50848195-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 07:22:35 crc kubenswrapper[4809]: I1124 07:22:35.693897 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nq4l\" (UniqueName: \"kubernetes.io/projected/8a942ac4-7652-45aa-b238-3aff50848195-kube-api-access-9nq4l\") on node \"crc\" DevicePath \"\"" Nov 24 07:22:36 crc kubenswrapper[4809]: I1124 07:22:36.027342 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k" event={"ID":"8a942ac4-7652-45aa-b238-3aff50848195","Type":"ContainerDied","Data":"4181142860d70def7bcb809f8a3ff53926958251d4faac3bfc775af86fbdb19a"} Nov 24 07:22:36 crc kubenswrapper[4809]: I1124 07:22:36.027393 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4181142860d70def7bcb809f8a3ff53926958251d4faac3bfc775af86fbdb19a" Nov 24 07:22:36 crc kubenswrapper[4809]: I1124 07:22:36.027402 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k" Nov 24 07:22:36 crc kubenswrapper[4809]: I1124 07:22:36.115755 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f"] Nov 24 07:22:36 crc kubenswrapper[4809]: E1124 07:22:36.116397 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a942ac4-7652-45aa-b238-3aff50848195" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 07:22:36 crc kubenswrapper[4809]: I1124 07:22:36.116426 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a942ac4-7652-45aa-b238-3aff50848195" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 07:22:36 crc kubenswrapper[4809]: I1124 07:22:36.116707 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a942ac4-7652-45aa-b238-3aff50848195" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 07:22:36 crc kubenswrapper[4809]: I1124 07:22:36.117729 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f" Nov 24 07:22:36 crc kubenswrapper[4809]: I1124 07:22:36.120548 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 07:22:36 crc kubenswrapper[4809]: I1124 07:22:36.120894 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 07:22:36 crc kubenswrapper[4809]: I1124 07:22:36.120929 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 07:22:36 crc kubenswrapper[4809]: I1124 07:22:36.121061 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lx7ll" Nov 24 07:22:36 crc kubenswrapper[4809]: I1124 07:22:36.127268 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f"] Nov 24 07:22:36 crc kubenswrapper[4809]: I1124 07:22:36.203413 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f\" (UID: \"2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f" Nov 24 07:22:36 crc kubenswrapper[4809]: I1124 07:22:36.203463 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f\" (UID: \"2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f" Nov 24 07:22:36 crc kubenswrapper[4809]: I1124 07:22:36.203492 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgvjp\" (UniqueName: \"kubernetes.io/projected/2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd-kube-api-access-dgvjp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f\" (UID: \"2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f" Nov 24 07:22:36 crc kubenswrapper[4809]: I1124 07:22:36.305383 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f\" (UID: \"2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f" Nov 24 07:22:36 crc kubenswrapper[4809]: I1124 07:22:36.305445 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f\" (UID: \"2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f" Nov 24 07:22:36 crc kubenswrapper[4809]: I1124 07:22:36.305475 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgvjp\" (UniqueName: \"kubernetes.io/projected/2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd-kube-api-access-dgvjp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f\" (UID: \"2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f" Nov 24 07:22:36 crc kubenswrapper[4809]: I1124 07:22:36.310825 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f\" (UID: \"2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f" Nov 24 07:22:36 crc kubenswrapper[4809]: I1124 07:22:36.313511 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f\" (UID: \"2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f" Nov 24 07:22:36 crc kubenswrapper[4809]: I1124 07:22:36.325043 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgvjp\" (UniqueName: \"kubernetes.io/projected/2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd-kube-api-access-dgvjp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f\" (UID: \"2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f" Nov 24 07:22:36 crc kubenswrapper[4809]: I1124 07:22:36.442306 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f" Nov 24 07:22:37 crc kubenswrapper[4809]: I1124 07:22:37.023776 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f"] Nov 24 07:22:38 crc kubenswrapper[4809]: I1124 07:22:38.076457 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f" event={"ID":"2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd","Type":"ContainerStarted","Data":"6e044e58049ca7bd00f62fdd993556de1ff8f8a75ecfb6a9e9f5079f4f51c7bb"} Nov 24 07:22:38 crc kubenswrapper[4809]: I1124 07:22:38.077038 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f" event={"ID":"2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd","Type":"ContainerStarted","Data":"e1076fb64fe26851e4ad57b5fee63443d4ecf6daabe81dc24d39abacda573116"} Nov 24 07:22:38 crc kubenswrapper[4809]: I1124 07:22:38.100818 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f" podStartSLOduration=1.570386756 podStartE2EDuration="2.100798138s" podCreationTimestamp="2025-11-24 07:22:36 +0000 UTC" firstStartedPulling="2025-11-24 07:22:37.033935054 +0000 UTC m=+1656.934526759" lastFinishedPulling="2025-11-24 07:22:37.564346436 +0000 UTC m=+1657.464938141" observedRunningTime="2025-11-24 07:22:38.096595487 +0000 UTC m=+1657.997187242" watchObservedRunningTime="2025-11-24 07:22:38.100798138 +0000 UTC m=+1658.001389863" Nov 24 07:22:43 crc kubenswrapper[4809]: I1124 07:22:43.127900 4809 generic.go:334] "Generic (PLEG): container finished" podID="2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd" containerID="6e044e58049ca7bd00f62fdd993556de1ff8f8a75ecfb6a9e9f5079f4f51c7bb" exitCode=0 Nov 24 07:22:43 crc kubenswrapper[4809]: I1124 07:22:43.127956 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f" event={"ID":"2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd","Type":"ContainerDied","Data":"6e044e58049ca7bd00f62fdd993556de1ff8f8a75ecfb6a9e9f5079f4f51c7bb"} Nov 24 07:22:44 crc kubenswrapper[4809]: I1124 07:22:44.511535 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f" Nov 24 07:22:44 crc kubenswrapper[4809]: I1124 07:22:44.571229 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd-inventory\") pod \"2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd\" (UID: \"2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd\") " Nov 24 07:22:44 crc kubenswrapper[4809]: I1124 07:22:44.571705 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd-ssh-key\") pod \"2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd\" (UID: \"2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd\") " Nov 24 07:22:44 crc kubenswrapper[4809]: I1124 07:22:44.571837 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgvjp\" (UniqueName: \"kubernetes.io/projected/2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd-kube-api-access-dgvjp\") pod \"2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd\" (UID: \"2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd\") " Nov 24 07:22:44 crc kubenswrapper[4809]: I1124 07:22:44.577897 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd-kube-api-access-dgvjp" (OuterVolumeSpecName: "kube-api-access-dgvjp") pod "2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd" (UID: "2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd"). InnerVolumeSpecName "kube-api-access-dgvjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:22:44 crc kubenswrapper[4809]: I1124 07:22:44.600200 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd" (UID: "2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:22:44 crc kubenswrapper[4809]: I1124 07:22:44.601826 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd-inventory" (OuterVolumeSpecName: "inventory") pod "2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd" (UID: "2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:22:44 crc kubenswrapper[4809]: I1124 07:22:44.674372 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 07:22:44 crc kubenswrapper[4809]: I1124 07:22:44.674412 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 07:22:44 crc kubenswrapper[4809]: I1124 07:22:44.674426 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgvjp\" (UniqueName: \"kubernetes.io/projected/2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd-kube-api-access-dgvjp\") on node \"crc\" DevicePath \"\"" Nov 24 07:22:45 crc kubenswrapper[4809]: I1124 07:22:45.147702 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f" event={"ID":"2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd","Type":"ContainerDied","Data":"e1076fb64fe26851e4ad57b5fee63443d4ecf6daabe81dc24d39abacda573116"} Nov 24 07:22:45 crc kubenswrapper[4809]: I1124 07:22:45.147750 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1076fb64fe26851e4ad57b5fee63443d4ecf6daabe81dc24d39abacda573116" Nov 24 07:22:45 crc kubenswrapper[4809]: I1124 07:22:45.147840 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f" Nov 24 07:22:45 crc kubenswrapper[4809]: I1124 07:22:45.249912 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-2fc9j"] Nov 24 07:22:45 crc kubenswrapper[4809]: E1124 07:22:45.250590 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 07:22:45 crc kubenswrapper[4809]: I1124 07:22:45.250627 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 07:22:45 crc kubenswrapper[4809]: I1124 07:22:45.251070 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 07:22:45 crc kubenswrapper[4809]: I1124 07:22:45.252054 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2fc9j" Nov 24 07:22:45 crc kubenswrapper[4809]: I1124 07:22:45.254871 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 07:22:45 crc kubenswrapper[4809]: I1124 07:22:45.255475 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lx7ll" Nov 24 07:22:45 crc kubenswrapper[4809]: I1124 07:22:45.256317 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 07:22:45 crc kubenswrapper[4809]: I1124 07:22:45.256620 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 07:22:45 crc kubenswrapper[4809]: I1124 07:22:45.285351 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/241ce393-aede-4973-adcf-93efc73bcd32-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2fc9j\" (UID: \"241ce393-aede-4973-adcf-93efc73bcd32\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2fc9j" Nov 24 07:22:45 crc kubenswrapper[4809]: I1124 07:22:45.285907 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sph67\" (UniqueName: \"kubernetes.io/projected/241ce393-aede-4973-adcf-93efc73bcd32-kube-api-access-sph67\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2fc9j\" (UID: \"241ce393-aede-4973-adcf-93efc73bcd32\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2fc9j" Nov 24 07:22:45 crc kubenswrapper[4809]: I1124 07:22:45.286357 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/241ce393-aede-4973-adcf-93efc73bcd32-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2fc9j\" (UID: \"241ce393-aede-4973-adcf-93efc73bcd32\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2fc9j" Nov 24 07:22:45 crc kubenswrapper[4809]: I1124 07:22:45.296954 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-2fc9j"] Nov 24 07:22:45 crc kubenswrapper[4809]: I1124 07:22:45.388520 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/241ce393-aede-4973-adcf-93efc73bcd32-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2fc9j\" (UID: \"241ce393-aede-4973-adcf-93efc73bcd32\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2fc9j" Nov 24 07:22:45 crc kubenswrapper[4809]: I1124 07:22:45.388600 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sph67\" (UniqueName: \"kubernetes.io/projected/241ce393-aede-4973-adcf-93efc73bcd32-kube-api-access-sph67\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2fc9j\" (UID: \"241ce393-aede-4973-adcf-93efc73bcd32\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2fc9j" Nov 24 07:22:45 crc kubenswrapper[4809]: I1124 07:22:45.388651 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/241ce393-aede-4973-adcf-93efc73bcd32-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2fc9j\" (UID: \"241ce393-aede-4973-adcf-93efc73bcd32\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2fc9j" Nov 24 07:22:45 crc kubenswrapper[4809]: I1124 07:22:45.391701 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/241ce393-aede-4973-adcf-93efc73bcd32-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2fc9j\" (UID: \"241ce393-aede-4973-adcf-93efc73bcd32\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2fc9j" Nov 24 07:22:45 crc kubenswrapper[4809]: I1124 07:22:45.392690 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/241ce393-aede-4973-adcf-93efc73bcd32-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2fc9j\" (UID: \"241ce393-aede-4973-adcf-93efc73bcd32\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2fc9j" Nov 24 07:22:45 crc kubenswrapper[4809]: I1124 07:22:45.411114 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sph67\" (UniqueName: \"kubernetes.io/projected/241ce393-aede-4973-adcf-93efc73bcd32-kube-api-access-sph67\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2fc9j\" (UID: \"241ce393-aede-4973-adcf-93efc73bcd32\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2fc9j" Nov 24 07:22:45 crc kubenswrapper[4809]: I1124 07:22:45.574500 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2fc9j" Nov 24 07:22:46 crc kubenswrapper[4809]: I1124 07:22:46.097698 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-2fc9j"] Nov 24 07:22:46 crc kubenswrapper[4809]: I1124 07:22:46.159949 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2fc9j" event={"ID":"241ce393-aede-4973-adcf-93efc73bcd32","Type":"ContainerStarted","Data":"ea8422b7a1800a02aee49ce9f34f49eb09c9210d9cb3efe7baf6db6f99383e65"} Nov 24 07:22:47 crc kubenswrapper[4809]: I1124 07:22:47.170929 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2fc9j" event={"ID":"241ce393-aede-4973-adcf-93efc73bcd32","Type":"ContainerStarted","Data":"9090dca0620d7ebc0484a0e48825fcc2582f12e8656dc0f3f44d7b5c4c545616"} Nov 24 07:22:47 crc kubenswrapper[4809]: I1124 07:22:47.189237 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2fc9j" podStartSLOduration=1.5676640210000001 podStartE2EDuration="2.1892191s" podCreationTimestamp="2025-11-24 07:22:45 +0000 UTC" firstStartedPulling="2025-11-24 07:22:46.107346686 +0000 UTC m=+1666.007938421" lastFinishedPulling="2025-11-24 07:22:46.728901785 +0000 UTC m=+1666.629493500" observedRunningTime="2025-11-24 07:22:47.183429306 +0000 UTC m=+1667.084021011" watchObservedRunningTime="2025-11-24 07:22:47.1892191 +0000 UTC m=+1667.089810795" Nov 24 07:22:48 crc kubenswrapper[4809]: I1124 07:22:48.891829 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:22:48 crc kubenswrapper[4809]: E1124 07:22:48.892701 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:23:02 crc kubenswrapper[4809]: I1124 07:23:02.891252 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:23:02 crc kubenswrapper[4809]: E1124 07:23:02.891926 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:23:06 crc kubenswrapper[4809]: I1124 07:23:06.066954 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-l7mxp"] Nov 24 07:23:06 crc kubenswrapper[4809]: I1124 07:23:06.079503 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-q8dsb"] Nov 24 07:23:06 crc kubenswrapper[4809]: I1124 07:23:06.086850 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-wn4xp"] Nov 24 07:23:06 crc kubenswrapper[4809]: I1124 07:23:06.093433 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-6e9f-account-create-7788v"] Nov 24 07:23:06 crc kubenswrapper[4809]: I1124 07:23:06.099755 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-l7mxp"] Nov 24 07:23:06 crc kubenswrapper[4809]: I1124 07:23:06.107002 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-q8dsb"] Nov 24 07:23:06 crc kubenswrapper[4809]: I1124 07:23:06.130804 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-f811-account-create-tt8dk"] Nov 24 07:23:06 crc kubenswrapper[4809]: I1124 07:23:06.136063 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-6e9f-account-create-7788v"] Nov 24 07:23:06 crc kubenswrapper[4809]: I1124 07:23:06.144606 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-f811-account-create-tt8dk"] Nov 24 07:23:06 crc kubenswrapper[4809]: I1124 07:23:06.151107 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-4079-account-create-7bgmp"] Nov 24 07:23:06 crc kubenswrapper[4809]: I1124 07:23:06.157435 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-4079-account-create-7bgmp"] Nov 24 07:23:06 crc kubenswrapper[4809]: I1124 07:23:06.164028 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-wn4xp"] Nov 24 07:23:06 crc kubenswrapper[4809]: I1124 07:23:06.903855 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab5f01-bc77-40d0-aacc-3af0e5c09e3f" path="/var/lib/kubelet/pods/01ab5f01-bc77-40d0-aacc-3af0e5c09e3f/volumes" Nov 24 07:23:06 crc kubenswrapper[4809]: I1124 07:23:06.905127 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25b3b27c-bf48-4067-b9fa-2312ff79fe66" path="/var/lib/kubelet/pods/25b3b27c-bf48-4067-b9fa-2312ff79fe66/volumes" Nov 24 07:23:06 crc kubenswrapper[4809]: I1124 07:23:06.906213 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4197d15d-b7aa-4efe-ad80-c3d731afbc44" path="/var/lib/kubelet/pods/4197d15d-b7aa-4efe-ad80-c3d731afbc44/volumes" Nov 24 07:23:06 crc kubenswrapper[4809]: I1124 07:23:06.907322 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="603e7b5e-01e3-4016-a1bb-9842c9e9ae1e" path="/var/lib/kubelet/pods/603e7b5e-01e3-4016-a1bb-9842c9e9ae1e/volumes" Nov 24 07:23:06 crc kubenswrapper[4809]: I1124 07:23:06.909469 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d43a2b0-ce7c-4392-a104-74b2a2a725c1" path="/var/lib/kubelet/pods/7d43a2b0-ce7c-4392-a104-74b2a2a725c1/volumes" Nov 24 07:23:06 crc kubenswrapper[4809]: I1124 07:23:06.910611 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5cfe52d-a6fb-4bf6-a08b-10f6c6547046" path="/var/lib/kubelet/pods/e5cfe52d-a6fb-4bf6-a08b-10f6c6547046/volumes" Nov 24 07:23:08 crc kubenswrapper[4809]: I1124 07:23:08.850552 4809 scope.go:117] "RemoveContainer" containerID="75d77f4aa40945e661f871d73ee743fffc6fcfa385d15c745bba59849623f594" Nov 24 07:23:08 crc kubenswrapper[4809]: I1124 07:23:08.944951 4809 scope.go:117] "RemoveContainer" containerID="2c1da7380ece890327fe29583a5580de0895e39d1b91fc06834a63134f70c22a" Nov 24 07:23:08 crc kubenswrapper[4809]: I1124 07:23:08.985785 4809 scope.go:117] "RemoveContainer" containerID="fe9aa5446cf96b6a630e785504668834e307a395a9d36f2a44ceb7c7db18e1c5" Nov 24 07:23:09 crc kubenswrapper[4809]: I1124 07:23:09.003259 4809 scope.go:117] "RemoveContainer" containerID="c0a14a8affcc77fa698259e63f590d8e16e0632574118a5aaa166805e9e9aeed" Nov 24 07:23:09 crc kubenswrapper[4809]: I1124 07:23:09.066560 4809 scope.go:117] "RemoveContainer" containerID="26141a9ca9d9f8e91477625f405b84918fcb3557393e3ba6e068fe5e60c7460f" Nov 24 07:23:09 crc kubenswrapper[4809]: I1124 07:23:09.088876 4809 scope.go:117] "RemoveContainer" containerID="0d7c52cf5c5cc03d5a34b01cf1a3e2ef1ec36b4324033c502207eeed4da969fa" Nov 24 07:23:09 crc kubenswrapper[4809]: I1124 07:23:09.134808 4809 scope.go:117] "RemoveContainer" containerID="e7647fa25d4f30aa0e004b6473877f7a10c062de7bbd7834170076f416f679cf" Nov 24 07:23:09 crc kubenswrapper[4809]: I1124 07:23:09.158168 4809 scope.go:117] "RemoveContainer" containerID="01ef821104ff2fbdbeaf37ceb28bf62950efaea3723a70ac8cfabc611a5820a4" Nov 24 07:23:09 crc kubenswrapper[4809]: I1124 07:23:09.198058 4809 scope.go:117] "RemoveContainer" containerID="5999e58fde6bd9838c96631ddabea81657a63f5690505976131497160a9369db" Nov 24 07:23:09 crc kubenswrapper[4809]: I1124 07:23:09.225372 4809 scope.go:117] "RemoveContainer" containerID="6b7f2215ad35ca949d1beba1c83c3a35b4b828adba3902249bf4be56100f2f4c" Nov 24 07:23:09 crc kubenswrapper[4809]: I1124 07:23:09.249301 4809 scope.go:117] "RemoveContainer" containerID="d15df5b7d8e85b6a2684a9dbf477405a616065e521d3d10b6c8595965f9577f4" Nov 24 07:23:16 crc kubenswrapper[4809]: I1124 07:23:16.892237 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:23:16 crc kubenswrapper[4809]: E1124 07:23:16.892991 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:23:30 crc kubenswrapper[4809]: I1124 07:23:30.034072 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-779kx"] Nov 24 07:23:30 crc kubenswrapper[4809]: I1124 07:23:30.041708 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-779kx"] Nov 24 07:23:30 crc kubenswrapper[4809]: I1124 07:23:30.904022 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6217b72-33d3-4b2e-86a2-10bd847003e4" path="/var/lib/kubelet/pods/b6217b72-33d3-4b2e-86a2-10bd847003e4/volumes" Nov 24 07:23:31 crc kubenswrapper[4809]: I1124 07:23:31.637572 4809 generic.go:334] "Generic (PLEG): container finished" podID="241ce393-aede-4973-adcf-93efc73bcd32" containerID="9090dca0620d7ebc0484a0e48825fcc2582f12e8656dc0f3f44d7b5c4c545616" exitCode=0 Nov 24 07:23:31 crc kubenswrapper[4809]: I1124 07:23:31.637620 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2fc9j" event={"ID":"241ce393-aede-4973-adcf-93efc73bcd32","Type":"ContainerDied","Data":"9090dca0620d7ebc0484a0e48825fcc2582f12e8656dc0f3f44d7b5c4c545616"} Nov 24 07:23:31 crc kubenswrapper[4809]: I1124 07:23:31.891300 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:23:31 crc kubenswrapper[4809]: E1124 07:23:31.891531 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.159915 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2fc9j" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.227362 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sph67\" (UniqueName: \"kubernetes.io/projected/241ce393-aede-4973-adcf-93efc73bcd32-kube-api-access-sph67\") pod \"241ce393-aede-4973-adcf-93efc73bcd32\" (UID: \"241ce393-aede-4973-adcf-93efc73bcd32\") " Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.227457 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/241ce393-aede-4973-adcf-93efc73bcd32-inventory\") pod \"241ce393-aede-4973-adcf-93efc73bcd32\" (UID: \"241ce393-aede-4973-adcf-93efc73bcd32\") " Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.227487 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/241ce393-aede-4973-adcf-93efc73bcd32-ssh-key\") pod \"241ce393-aede-4973-adcf-93efc73bcd32\" (UID: \"241ce393-aede-4973-adcf-93efc73bcd32\") " Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.233544 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/241ce393-aede-4973-adcf-93efc73bcd32-kube-api-access-sph67" (OuterVolumeSpecName: "kube-api-access-sph67") pod "241ce393-aede-4973-adcf-93efc73bcd32" (UID: "241ce393-aede-4973-adcf-93efc73bcd32"). InnerVolumeSpecName "kube-api-access-sph67". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.253946 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/241ce393-aede-4973-adcf-93efc73bcd32-inventory" (OuterVolumeSpecName: "inventory") pod "241ce393-aede-4973-adcf-93efc73bcd32" (UID: "241ce393-aede-4973-adcf-93efc73bcd32"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.258184 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/241ce393-aede-4973-adcf-93efc73bcd32-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "241ce393-aede-4973-adcf-93efc73bcd32" (UID: "241ce393-aede-4973-adcf-93efc73bcd32"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.330158 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/241ce393-aede-4973-adcf-93efc73bcd32-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.330210 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/241ce393-aede-4973-adcf-93efc73bcd32-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.330231 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sph67\" (UniqueName: \"kubernetes.io/projected/241ce393-aede-4973-adcf-93efc73bcd32-kube-api-access-sph67\") on node \"crc\" DevicePath \"\"" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.661939 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2fc9j" event={"ID":"241ce393-aede-4973-adcf-93efc73bcd32","Type":"ContainerDied","Data":"ea8422b7a1800a02aee49ce9f34f49eb09c9210d9cb3efe7baf6db6f99383e65"} Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.662047 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea8422b7a1800a02aee49ce9f34f49eb09c9210d9cb3efe7baf6db6f99383e65" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.662170 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2fc9j" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.766336 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr"] Nov 24 07:23:33 crc kubenswrapper[4809]: E1124 07:23:33.766828 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="241ce393-aede-4973-adcf-93efc73bcd32" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.766854 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="241ce393-aede-4973-adcf-93efc73bcd32" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.767142 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="241ce393-aede-4973-adcf-93efc73bcd32" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.767929 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.771027 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.771180 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.771779 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lx7ll" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.773246 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.778018 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr"] Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.839710 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7f9b89a-871e-4478-bb01-1120cd72a2b6-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr\" (UID: \"a7f9b89a-871e-4478-bb01-1120cd72a2b6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.840051 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh4dx\" (UniqueName: \"kubernetes.io/projected/a7f9b89a-871e-4478-bb01-1120cd72a2b6-kube-api-access-xh4dx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr\" (UID: \"a7f9b89a-871e-4478-bb01-1120cd72a2b6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.840422 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7f9b89a-871e-4478-bb01-1120cd72a2b6-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr\" (UID: \"a7f9b89a-871e-4478-bb01-1120cd72a2b6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.941892 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7f9b89a-871e-4478-bb01-1120cd72a2b6-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr\" (UID: \"a7f9b89a-871e-4478-bb01-1120cd72a2b6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.942015 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh4dx\" (UniqueName: \"kubernetes.io/projected/a7f9b89a-871e-4478-bb01-1120cd72a2b6-kube-api-access-xh4dx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr\" (UID: \"a7f9b89a-871e-4478-bb01-1120cd72a2b6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.942102 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7f9b89a-871e-4478-bb01-1120cd72a2b6-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr\" (UID: \"a7f9b89a-871e-4478-bb01-1120cd72a2b6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.952760 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7f9b89a-871e-4478-bb01-1120cd72a2b6-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr\" (UID: \"a7f9b89a-871e-4478-bb01-1120cd72a2b6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.952776 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7f9b89a-871e-4478-bb01-1120cd72a2b6-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr\" (UID: \"a7f9b89a-871e-4478-bb01-1120cd72a2b6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr" Nov 24 07:23:33 crc kubenswrapper[4809]: I1124 07:23:33.960364 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh4dx\" (UniqueName: \"kubernetes.io/projected/a7f9b89a-871e-4478-bb01-1120cd72a2b6-kube-api-access-xh4dx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr\" (UID: \"a7f9b89a-871e-4478-bb01-1120cd72a2b6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr" Nov 24 07:23:34 crc kubenswrapper[4809]: I1124 07:23:34.095060 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr" Nov 24 07:23:34 crc kubenswrapper[4809]: I1124 07:23:34.468496 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr"] Nov 24 07:23:34 crc kubenswrapper[4809]: I1124 07:23:34.671366 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr" event={"ID":"a7f9b89a-871e-4478-bb01-1120cd72a2b6","Type":"ContainerStarted","Data":"e219f8fd90d4b5e17e41340a453913653f693d854d24677de27faa99496d3431"} Nov 24 07:23:35 crc kubenswrapper[4809]: I1124 07:23:35.684072 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr" event={"ID":"a7f9b89a-871e-4478-bb01-1120cd72a2b6","Type":"ContainerStarted","Data":"234c91ebcc433537e0e9f969a0176c1b362d93111062b88568f6db5758df74d9"} Nov 24 07:23:35 crc kubenswrapper[4809]: I1124 07:23:35.704789 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr" podStartSLOduration=2.279702521 podStartE2EDuration="2.704768247s" podCreationTimestamp="2025-11-24 07:23:33 +0000 UTC" firstStartedPulling="2025-11-24 07:23:34.476931393 +0000 UTC m=+1714.377523108" lastFinishedPulling="2025-11-24 07:23:34.901997139 +0000 UTC m=+1714.802588834" observedRunningTime="2025-11-24 07:23:35.699503066 +0000 UTC m=+1715.600094811" watchObservedRunningTime="2025-11-24 07:23:35.704768247 +0000 UTC m=+1715.605359962" Nov 24 07:23:42 crc kubenswrapper[4809]: I1124 07:23:42.892167 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:23:42 crc kubenswrapper[4809]: E1124 07:23:42.893234 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:23:51 crc kubenswrapper[4809]: I1124 07:23:51.036044 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-8wwqj"] Nov 24 07:23:51 crc kubenswrapper[4809]: I1124 07:23:51.050642 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-8wwqj"] Nov 24 07:23:52 crc kubenswrapper[4809]: I1124 07:23:52.904568 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="825d5761-5b87-4e12-8e7c-392e252331e0" path="/var/lib/kubelet/pods/825d5761-5b87-4e12-8e7c-392e252331e0/volumes" Nov 24 07:23:53 crc kubenswrapper[4809]: I1124 07:23:53.039335 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9fqbw"] Nov 24 07:23:53 crc kubenswrapper[4809]: I1124 07:23:53.046651 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9fqbw"] Nov 24 07:23:54 crc kubenswrapper[4809]: I1124 07:23:54.909808 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45003542-95f6-46b1-8ff9-06dfc99c1d93" path="/var/lib/kubelet/pods/45003542-95f6-46b1-8ff9-06dfc99c1d93/volumes" Nov 24 07:23:57 crc kubenswrapper[4809]: I1124 07:23:57.892159 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:23:57 crc kubenswrapper[4809]: E1124 07:23:57.893200 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:24:09 crc kubenswrapper[4809]: I1124 07:24:09.541473 4809 scope.go:117] "RemoveContainer" containerID="c02fa30371d4010318a10fd542ad941354969f99e6d49d8655b2cc7d516dc5ed" Nov 24 07:24:09 crc kubenswrapper[4809]: I1124 07:24:09.596304 4809 scope.go:117] "RemoveContainer" containerID="94fd2308eff927bad656d650cf3af2eb5d33358e78b1e4aaeac0735a4459c80c" Nov 24 07:24:09 crc kubenswrapper[4809]: I1124 07:24:09.672371 4809 scope.go:117] "RemoveContainer" containerID="af048ce9879c779790b2123c8a7427bdd8f5606ae5dac57293e6c491a0d84d8e" Nov 24 07:24:11 crc kubenswrapper[4809]: I1124 07:24:11.891604 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:24:11 crc kubenswrapper[4809]: E1124 07:24:11.892228 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:24:26 crc kubenswrapper[4809]: I1124 07:24:26.891390 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:24:26 crc kubenswrapper[4809]: E1124 07:24:26.892524 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:24:33 crc kubenswrapper[4809]: I1124 07:24:33.257135 4809 generic.go:334] "Generic (PLEG): container finished" podID="a7f9b89a-871e-4478-bb01-1120cd72a2b6" containerID="234c91ebcc433537e0e9f969a0176c1b362d93111062b88568f6db5758df74d9" exitCode=0 Nov 24 07:24:33 crc kubenswrapper[4809]: I1124 07:24:33.257226 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr" event={"ID":"a7f9b89a-871e-4478-bb01-1120cd72a2b6","Type":"ContainerDied","Data":"234c91ebcc433537e0e9f969a0176c1b362d93111062b88568f6db5758df74d9"} Nov 24 07:24:34 crc kubenswrapper[4809]: I1124 07:24:34.632001 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr" Nov 24 07:24:34 crc kubenswrapper[4809]: I1124 07:24:34.725619 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xh4dx\" (UniqueName: \"kubernetes.io/projected/a7f9b89a-871e-4478-bb01-1120cd72a2b6-kube-api-access-xh4dx\") pod \"a7f9b89a-871e-4478-bb01-1120cd72a2b6\" (UID: \"a7f9b89a-871e-4478-bb01-1120cd72a2b6\") " Nov 24 07:24:34 crc kubenswrapper[4809]: I1124 07:24:34.725737 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7f9b89a-871e-4478-bb01-1120cd72a2b6-inventory\") pod \"a7f9b89a-871e-4478-bb01-1120cd72a2b6\" (UID: \"a7f9b89a-871e-4478-bb01-1120cd72a2b6\") " Nov 24 07:24:34 crc kubenswrapper[4809]: I1124 07:24:34.725782 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7f9b89a-871e-4478-bb01-1120cd72a2b6-ssh-key\") pod \"a7f9b89a-871e-4478-bb01-1120cd72a2b6\" (UID: \"a7f9b89a-871e-4478-bb01-1120cd72a2b6\") " Nov 24 07:24:34 crc kubenswrapper[4809]: I1124 07:24:34.730770 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7f9b89a-871e-4478-bb01-1120cd72a2b6-kube-api-access-xh4dx" (OuterVolumeSpecName: "kube-api-access-xh4dx") pod "a7f9b89a-871e-4478-bb01-1120cd72a2b6" (UID: "a7f9b89a-871e-4478-bb01-1120cd72a2b6"). InnerVolumeSpecName "kube-api-access-xh4dx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:24:34 crc kubenswrapper[4809]: I1124 07:24:34.750762 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7f9b89a-871e-4478-bb01-1120cd72a2b6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a7f9b89a-871e-4478-bb01-1120cd72a2b6" (UID: "a7f9b89a-871e-4478-bb01-1120cd72a2b6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:24:34 crc kubenswrapper[4809]: I1124 07:24:34.769583 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7f9b89a-871e-4478-bb01-1120cd72a2b6-inventory" (OuterVolumeSpecName: "inventory") pod "a7f9b89a-871e-4478-bb01-1120cd72a2b6" (UID: "a7f9b89a-871e-4478-bb01-1120cd72a2b6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:24:34 crc kubenswrapper[4809]: I1124 07:24:34.828470 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xh4dx\" (UniqueName: \"kubernetes.io/projected/a7f9b89a-871e-4478-bb01-1120cd72a2b6-kube-api-access-xh4dx\") on node \"crc\" DevicePath \"\"" Nov 24 07:24:34 crc kubenswrapper[4809]: I1124 07:24:34.828505 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7f9b89a-871e-4478-bb01-1120cd72a2b6-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 07:24:34 crc kubenswrapper[4809]: I1124 07:24:34.828518 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7f9b89a-871e-4478-bb01-1120cd72a2b6-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 07:24:35 crc kubenswrapper[4809]: I1124 07:24:35.278225 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr" event={"ID":"a7f9b89a-871e-4478-bb01-1120cd72a2b6","Type":"ContainerDied","Data":"e219f8fd90d4b5e17e41340a453913653f693d854d24677de27faa99496d3431"} Nov 24 07:24:35 crc kubenswrapper[4809]: I1124 07:24:35.278275 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr" Nov 24 07:24:35 crc kubenswrapper[4809]: I1124 07:24:35.278289 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e219f8fd90d4b5e17e41340a453913653f693d854d24677de27faa99496d3431" Nov 24 07:24:35 crc kubenswrapper[4809]: I1124 07:24:35.439285 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ckss4"] Nov 24 07:24:35 crc kubenswrapper[4809]: E1124 07:24:35.439782 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f9b89a-871e-4478-bb01-1120cd72a2b6" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 07:24:35 crc kubenswrapper[4809]: I1124 07:24:35.439803 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f9b89a-871e-4478-bb01-1120cd72a2b6" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 07:24:35 crc kubenswrapper[4809]: I1124 07:24:35.440066 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7f9b89a-871e-4478-bb01-1120cd72a2b6" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 07:24:35 crc kubenswrapper[4809]: I1124 07:24:35.440868 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ckss4" Nov 24 07:24:35 crc kubenswrapper[4809]: I1124 07:24:35.442659 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lx7ll" Nov 24 07:24:35 crc kubenswrapper[4809]: I1124 07:24:35.443562 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 07:24:35 crc kubenswrapper[4809]: I1124 07:24:35.443791 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 07:24:35 crc kubenswrapper[4809]: I1124 07:24:35.444165 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 07:24:35 crc kubenswrapper[4809]: I1124 07:24:35.460008 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ckss4"] Nov 24 07:24:35 crc kubenswrapper[4809]: I1124 07:24:35.539898 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/801a6c69-8f03-4f88-ba22-cd68875e8ae2-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ckss4\" (UID: \"801a6c69-8f03-4f88-ba22-cd68875e8ae2\") " pod="openstack/ssh-known-hosts-edpm-deployment-ckss4" Nov 24 07:24:35 crc kubenswrapper[4809]: I1124 07:24:35.540047 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6n26\" (UniqueName: \"kubernetes.io/projected/801a6c69-8f03-4f88-ba22-cd68875e8ae2-kube-api-access-x6n26\") pod \"ssh-known-hosts-edpm-deployment-ckss4\" (UID: \"801a6c69-8f03-4f88-ba22-cd68875e8ae2\") " pod="openstack/ssh-known-hosts-edpm-deployment-ckss4" Nov 24 07:24:35 crc kubenswrapper[4809]: I1124 07:24:35.540079 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/801a6c69-8f03-4f88-ba22-cd68875e8ae2-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ckss4\" (UID: \"801a6c69-8f03-4f88-ba22-cd68875e8ae2\") " pod="openstack/ssh-known-hosts-edpm-deployment-ckss4" Nov 24 07:24:35 crc kubenswrapper[4809]: I1124 07:24:35.642192 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/801a6c69-8f03-4f88-ba22-cd68875e8ae2-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ckss4\" (UID: \"801a6c69-8f03-4f88-ba22-cd68875e8ae2\") " pod="openstack/ssh-known-hosts-edpm-deployment-ckss4" Nov 24 07:24:35 crc kubenswrapper[4809]: I1124 07:24:35.642415 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6n26\" (UniqueName: \"kubernetes.io/projected/801a6c69-8f03-4f88-ba22-cd68875e8ae2-kube-api-access-x6n26\") pod \"ssh-known-hosts-edpm-deployment-ckss4\" (UID: \"801a6c69-8f03-4f88-ba22-cd68875e8ae2\") " pod="openstack/ssh-known-hosts-edpm-deployment-ckss4" Nov 24 07:24:35 crc kubenswrapper[4809]: I1124 07:24:35.642465 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/801a6c69-8f03-4f88-ba22-cd68875e8ae2-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ckss4\" (UID: \"801a6c69-8f03-4f88-ba22-cd68875e8ae2\") " pod="openstack/ssh-known-hosts-edpm-deployment-ckss4" Nov 24 07:24:35 crc kubenswrapper[4809]: I1124 07:24:35.648633 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/801a6c69-8f03-4f88-ba22-cd68875e8ae2-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ckss4\" (UID: \"801a6c69-8f03-4f88-ba22-cd68875e8ae2\") " pod="openstack/ssh-known-hosts-edpm-deployment-ckss4" Nov 24 07:24:35 crc kubenswrapper[4809]: I1124 07:24:35.648721 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/801a6c69-8f03-4f88-ba22-cd68875e8ae2-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ckss4\" (UID: \"801a6c69-8f03-4f88-ba22-cd68875e8ae2\") " pod="openstack/ssh-known-hosts-edpm-deployment-ckss4" Nov 24 07:24:35 crc kubenswrapper[4809]: I1124 07:24:35.671738 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6n26\" (UniqueName: \"kubernetes.io/projected/801a6c69-8f03-4f88-ba22-cd68875e8ae2-kube-api-access-x6n26\") pod \"ssh-known-hosts-edpm-deployment-ckss4\" (UID: \"801a6c69-8f03-4f88-ba22-cd68875e8ae2\") " pod="openstack/ssh-known-hosts-edpm-deployment-ckss4" Nov 24 07:24:35 crc kubenswrapper[4809]: I1124 07:24:35.813563 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ckss4" Nov 24 07:24:36 crc kubenswrapper[4809]: I1124 07:24:36.379555 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ckss4"] Nov 24 07:24:37 crc kubenswrapper[4809]: I1124 07:24:37.062024 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-9v7n9"] Nov 24 07:24:37 crc kubenswrapper[4809]: I1124 07:24:37.069418 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-9v7n9"] Nov 24 07:24:37 crc kubenswrapper[4809]: I1124 07:24:37.297487 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ckss4" event={"ID":"801a6c69-8f03-4f88-ba22-cd68875e8ae2","Type":"ContainerStarted","Data":"0f2bcb535f546b065ed48f7d150789e7ffb64e7cd3137c63350e06e014f4747a"} Nov 24 07:24:37 crc kubenswrapper[4809]: I1124 07:24:37.297768 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ckss4" event={"ID":"801a6c69-8f03-4f88-ba22-cd68875e8ae2","Type":"ContainerStarted","Data":"36e8738d45fb7afbf587868bc83581c8b0e66aaaf2c7dcf910e26c3bc853c1e1"} Nov 24 07:24:37 crc kubenswrapper[4809]: I1124 07:24:37.321356 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-ckss4" podStartSLOduration=1.85098841 podStartE2EDuration="2.321334594s" podCreationTimestamp="2025-11-24 07:24:35 +0000 UTC" firstStartedPulling="2025-11-24 07:24:36.386189884 +0000 UTC m=+1776.286781589" lastFinishedPulling="2025-11-24 07:24:36.856536068 +0000 UTC m=+1776.757127773" observedRunningTime="2025-11-24 07:24:37.308235334 +0000 UTC m=+1777.208827049" watchObservedRunningTime="2025-11-24 07:24:37.321334594 +0000 UTC m=+1777.221926309" Nov 24 07:24:38 crc kubenswrapper[4809]: I1124 07:24:38.906922 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97266dae-027e-469f-8934-298e38fbde54" path="/var/lib/kubelet/pods/97266dae-027e-469f-8934-298e38fbde54/volumes" Nov 24 07:24:39 crc kubenswrapper[4809]: I1124 07:24:39.891527 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:24:39 crc kubenswrapper[4809]: E1124 07:24:39.892152 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:24:44 crc kubenswrapper[4809]: I1124 07:24:44.361011 4809 generic.go:334] "Generic (PLEG): container finished" podID="801a6c69-8f03-4f88-ba22-cd68875e8ae2" containerID="0f2bcb535f546b065ed48f7d150789e7ffb64e7cd3137c63350e06e014f4747a" exitCode=0 Nov 24 07:24:44 crc kubenswrapper[4809]: I1124 07:24:44.361071 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ckss4" event={"ID":"801a6c69-8f03-4f88-ba22-cd68875e8ae2","Type":"ContainerDied","Data":"0f2bcb535f546b065ed48f7d150789e7ffb64e7cd3137c63350e06e014f4747a"} Nov 24 07:24:45 crc kubenswrapper[4809]: I1124 07:24:45.835423 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ckss4" Nov 24 07:24:45 crc kubenswrapper[4809]: I1124 07:24:45.935297 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/801a6c69-8f03-4f88-ba22-cd68875e8ae2-inventory-0\") pod \"801a6c69-8f03-4f88-ba22-cd68875e8ae2\" (UID: \"801a6c69-8f03-4f88-ba22-cd68875e8ae2\") " Nov 24 07:24:45 crc kubenswrapper[4809]: I1124 07:24:45.935391 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6n26\" (UniqueName: \"kubernetes.io/projected/801a6c69-8f03-4f88-ba22-cd68875e8ae2-kube-api-access-x6n26\") pod \"801a6c69-8f03-4f88-ba22-cd68875e8ae2\" (UID: \"801a6c69-8f03-4f88-ba22-cd68875e8ae2\") " Nov 24 07:24:45 crc kubenswrapper[4809]: I1124 07:24:45.935550 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/801a6c69-8f03-4f88-ba22-cd68875e8ae2-ssh-key-openstack-edpm-ipam\") pod \"801a6c69-8f03-4f88-ba22-cd68875e8ae2\" (UID: \"801a6c69-8f03-4f88-ba22-cd68875e8ae2\") " Nov 24 07:24:45 crc kubenswrapper[4809]: I1124 07:24:45.948307 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/801a6c69-8f03-4f88-ba22-cd68875e8ae2-kube-api-access-x6n26" (OuterVolumeSpecName: "kube-api-access-x6n26") pod "801a6c69-8f03-4f88-ba22-cd68875e8ae2" (UID: "801a6c69-8f03-4f88-ba22-cd68875e8ae2"). InnerVolumeSpecName "kube-api-access-x6n26". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:24:45 crc kubenswrapper[4809]: I1124 07:24:45.966373 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/801a6c69-8f03-4f88-ba22-cd68875e8ae2-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "801a6c69-8f03-4f88-ba22-cd68875e8ae2" (UID: "801a6c69-8f03-4f88-ba22-cd68875e8ae2"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:24:45 crc kubenswrapper[4809]: I1124 07:24:45.976650 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/801a6c69-8f03-4f88-ba22-cd68875e8ae2-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "801a6c69-8f03-4f88-ba22-cd68875e8ae2" (UID: "801a6c69-8f03-4f88-ba22-cd68875e8ae2"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.039227 4809 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/801a6c69-8f03-4f88-ba22-cd68875e8ae2-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.039646 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6n26\" (UniqueName: \"kubernetes.io/projected/801a6c69-8f03-4f88-ba22-cd68875e8ae2-kube-api-access-x6n26\") on node \"crc\" DevicePath \"\"" Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.039665 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/801a6c69-8f03-4f88-ba22-cd68875e8ae2-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.385780 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ckss4" event={"ID":"801a6c69-8f03-4f88-ba22-cd68875e8ae2","Type":"ContainerDied","Data":"36e8738d45fb7afbf587868bc83581c8b0e66aaaf2c7dcf910e26c3bc853c1e1"} Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.385818 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36e8738d45fb7afbf587868bc83581c8b0e66aaaf2c7dcf910e26c3bc853c1e1" Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.385869 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ckss4" Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.462187 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-2fc45"] Nov 24 07:24:46 crc kubenswrapper[4809]: E1124 07:24:46.462601 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="801a6c69-8f03-4f88-ba22-cd68875e8ae2" containerName="ssh-known-hosts-edpm-deployment" Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.462626 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="801a6c69-8f03-4f88-ba22-cd68875e8ae2" containerName="ssh-known-hosts-edpm-deployment" Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.462860 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="801a6c69-8f03-4f88-ba22-cd68875e8ae2" containerName="ssh-known-hosts-edpm-deployment" Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.463559 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2fc45" Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.466525 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.467708 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.467961 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lx7ll" Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.468945 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.496806 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-2fc45"] Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.548459 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a469d7b-c3d8-46cf-b6cf-7fe13619f922-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2fc45\" (UID: \"2a469d7b-c3d8-46cf-b6cf-7fe13619f922\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2fc45" Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.548525 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a469d7b-c3d8-46cf-b6cf-7fe13619f922-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2fc45\" (UID: \"2a469d7b-c3d8-46cf-b6cf-7fe13619f922\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2fc45" Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.548555 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66snj\" (UniqueName: \"kubernetes.io/projected/2a469d7b-c3d8-46cf-b6cf-7fe13619f922-kube-api-access-66snj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2fc45\" (UID: \"2a469d7b-c3d8-46cf-b6cf-7fe13619f922\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2fc45" Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.650938 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a469d7b-c3d8-46cf-b6cf-7fe13619f922-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2fc45\" (UID: \"2a469d7b-c3d8-46cf-b6cf-7fe13619f922\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2fc45" Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.651121 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a469d7b-c3d8-46cf-b6cf-7fe13619f922-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2fc45\" (UID: \"2a469d7b-c3d8-46cf-b6cf-7fe13619f922\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2fc45" Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.651166 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66snj\" (UniqueName: \"kubernetes.io/projected/2a469d7b-c3d8-46cf-b6cf-7fe13619f922-kube-api-access-66snj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2fc45\" (UID: \"2a469d7b-c3d8-46cf-b6cf-7fe13619f922\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2fc45" Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.655177 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a469d7b-c3d8-46cf-b6cf-7fe13619f922-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2fc45\" (UID: \"2a469d7b-c3d8-46cf-b6cf-7fe13619f922\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2fc45" Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.656659 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a469d7b-c3d8-46cf-b6cf-7fe13619f922-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2fc45\" (UID: \"2a469d7b-c3d8-46cf-b6cf-7fe13619f922\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2fc45" Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.675785 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66snj\" (UniqueName: \"kubernetes.io/projected/2a469d7b-c3d8-46cf-b6cf-7fe13619f922-kube-api-access-66snj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2fc45\" (UID: \"2a469d7b-c3d8-46cf-b6cf-7fe13619f922\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2fc45" Nov 24 07:24:46 crc kubenswrapper[4809]: I1124 07:24:46.782443 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2fc45" Nov 24 07:24:47 crc kubenswrapper[4809]: I1124 07:24:47.357320 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-2fc45"] Nov 24 07:24:47 crc kubenswrapper[4809]: I1124 07:24:47.395000 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2fc45" event={"ID":"2a469d7b-c3d8-46cf-b6cf-7fe13619f922","Type":"ContainerStarted","Data":"360f87f1a1bddea22d8f0c407729a1898ca6ceb9da40e5e7887a01b15c7b66ad"} Nov 24 07:24:48 crc kubenswrapper[4809]: I1124 07:24:48.403845 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2fc45" event={"ID":"2a469d7b-c3d8-46cf-b6cf-7fe13619f922","Type":"ContainerStarted","Data":"d042d7d6fd3e4a4a8c28ae8694634f9d54102584bf0aea5e51428171246bd794"} Nov 24 07:24:48 crc kubenswrapper[4809]: I1124 07:24:48.435129 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2fc45" podStartSLOduration=2.001862109 podStartE2EDuration="2.435110921s" podCreationTimestamp="2025-11-24 07:24:46 +0000 UTC" firstStartedPulling="2025-11-24 07:24:47.357070175 +0000 UTC m=+1787.257661880" lastFinishedPulling="2025-11-24 07:24:47.790318967 +0000 UTC m=+1787.690910692" observedRunningTime="2025-11-24 07:24:48.428733721 +0000 UTC m=+1788.329325426" watchObservedRunningTime="2025-11-24 07:24:48.435110921 +0000 UTC m=+1788.335702616" Nov 24 07:24:52 crc kubenswrapper[4809]: I1124 07:24:52.891734 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:24:53 crc kubenswrapper[4809]: I1124 07:24:53.452731 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerStarted","Data":"da763efb03a569decc0792fe26fc2e80b6b2ae304146cd9bddaf55c16474daac"} Nov 24 07:24:57 crc kubenswrapper[4809]: I1124 07:24:57.490537 4809 generic.go:334] "Generic (PLEG): container finished" podID="2a469d7b-c3d8-46cf-b6cf-7fe13619f922" containerID="d042d7d6fd3e4a4a8c28ae8694634f9d54102584bf0aea5e51428171246bd794" exitCode=0 Nov 24 07:24:57 crc kubenswrapper[4809]: I1124 07:24:57.490642 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2fc45" event={"ID":"2a469d7b-c3d8-46cf-b6cf-7fe13619f922","Type":"ContainerDied","Data":"d042d7d6fd3e4a4a8c28ae8694634f9d54102584bf0aea5e51428171246bd794"} Nov 24 07:24:58 crc kubenswrapper[4809]: I1124 07:24:58.950885 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2fc45" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.018748 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a469d7b-c3d8-46cf-b6cf-7fe13619f922-ssh-key\") pod \"2a469d7b-c3d8-46cf-b6cf-7fe13619f922\" (UID: \"2a469d7b-c3d8-46cf-b6cf-7fe13619f922\") " Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.018827 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a469d7b-c3d8-46cf-b6cf-7fe13619f922-inventory\") pod \"2a469d7b-c3d8-46cf-b6cf-7fe13619f922\" (UID: \"2a469d7b-c3d8-46cf-b6cf-7fe13619f922\") " Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.019038 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66snj\" (UniqueName: \"kubernetes.io/projected/2a469d7b-c3d8-46cf-b6cf-7fe13619f922-kube-api-access-66snj\") pod \"2a469d7b-c3d8-46cf-b6cf-7fe13619f922\" (UID: \"2a469d7b-c3d8-46cf-b6cf-7fe13619f922\") " Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.024643 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a469d7b-c3d8-46cf-b6cf-7fe13619f922-kube-api-access-66snj" (OuterVolumeSpecName: "kube-api-access-66snj") pod "2a469d7b-c3d8-46cf-b6cf-7fe13619f922" (UID: "2a469d7b-c3d8-46cf-b6cf-7fe13619f922"). InnerVolumeSpecName "kube-api-access-66snj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.044534 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a469d7b-c3d8-46cf-b6cf-7fe13619f922-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2a469d7b-c3d8-46cf-b6cf-7fe13619f922" (UID: "2a469d7b-c3d8-46cf-b6cf-7fe13619f922"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.052297 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a469d7b-c3d8-46cf-b6cf-7fe13619f922-inventory" (OuterVolumeSpecName: "inventory") pod "2a469d7b-c3d8-46cf-b6cf-7fe13619f922" (UID: "2a469d7b-c3d8-46cf-b6cf-7fe13619f922"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.120509 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66snj\" (UniqueName: \"kubernetes.io/projected/2a469d7b-c3d8-46cf-b6cf-7fe13619f922-kube-api-access-66snj\") on node \"crc\" DevicePath \"\"" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.120558 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a469d7b-c3d8-46cf-b6cf-7fe13619f922-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.120578 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a469d7b-c3d8-46cf-b6cf-7fe13619f922-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.515621 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2fc45" event={"ID":"2a469d7b-c3d8-46cf-b6cf-7fe13619f922","Type":"ContainerDied","Data":"360f87f1a1bddea22d8f0c407729a1898ca6ceb9da40e5e7887a01b15c7b66ad"} Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.515678 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="360f87f1a1bddea22d8f0c407729a1898ca6ceb9da40e5e7887a01b15c7b66ad" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.515702 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2fc45" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.608693 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2"] Nov 24 07:24:59 crc kubenswrapper[4809]: E1124 07:24:59.609517 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a469d7b-c3d8-46cf-b6cf-7fe13619f922" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.609565 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a469d7b-c3d8-46cf-b6cf-7fe13619f922" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.610079 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a469d7b-c3d8-46cf-b6cf-7fe13619f922" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.611299 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.616870 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.617004 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.618149 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.618360 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lx7ll" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.635205 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a46047d8-9af2-4172-8e55-26dd29b587aa-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2\" (UID: \"a46047d8-9af2-4172-8e55-26dd29b587aa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.635394 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvhzg\" (UniqueName: \"kubernetes.io/projected/a46047d8-9af2-4172-8e55-26dd29b587aa-kube-api-access-dvhzg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2\" (UID: \"a46047d8-9af2-4172-8e55-26dd29b587aa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.635576 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a46047d8-9af2-4172-8e55-26dd29b587aa-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2\" (UID: \"a46047d8-9af2-4172-8e55-26dd29b587aa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.644483 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2"] Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.737112 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvhzg\" (UniqueName: \"kubernetes.io/projected/a46047d8-9af2-4172-8e55-26dd29b587aa-kube-api-access-dvhzg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2\" (UID: \"a46047d8-9af2-4172-8e55-26dd29b587aa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.737392 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a46047d8-9af2-4172-8e55-26dd29b587aa-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2\" (UID: \"a46047d8-9af2-4172-8e55-26dd29b587aa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.737712 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a46047d8-9af2-4172-8e55-26dd29b587aa-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2\" (UID: \"a46047d8-9af2-4172-8e55-26dd29b587aa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.740897 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a46047d8-9af2-4172-8e55-26dd29b587aa-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2\" (UID: \"a46047d8-9af2-4172-8e55-26dd29b587aa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.742710 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a46047d8-9af2-4172-8e55-26dd29b587aa-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2\" (UID: \"a46047d8-9af2-4172-8e55-26dd29b587aa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.756815 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvhzg\" (UniqueName: \"kubernetes.io/projected/a46047d8-9af2-4172-8e55-26dd29b587aa-kube-api-access-dvhzg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2\" (UID: \"a46047d8-9af2-4172-8e55-26dd29b587aa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2" Nov 24 07:24:59 crc kubenswrapper[4809]: I1124 07:24:59.938542 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2" Nov 24 07:25:00 crc kubenswrapper[4809]: I1124 07:25:00.494446 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2"] Nov 24 07:25:00 crc kubenswrapper[4809]: W1124 07:25:00.503912 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda46047d8_9af2_4172_8e55_26dd29b587aa.slice/crio-1b86aa996fe6cd146225df4c1be8378d06f4d46fde1036d06fd4b41468a6e44d WatchSource:0}: Error finding container 1b86aa996fe6cd146225df4c1be8378d06f4d46fde1036d06fd4b41468a6e44d: Status 404 returned error can't find the container with id 1b86aa996fe6cd146225df4c1be8378d06f4d46fde1036d06fd4b41468a6e44d Nov 24 07:25:00 crc kubenswrapper[4809]: I1124 07:25:00.528682 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2" event={"ID":"a46047d8-9af2-4172-8e55-26dd29b587aa","Type":"ContainerStarted","Data":"1b86aa996fe6cd146225df4c1be8378d06f4d46fde1036d06fd4b41468a6e44d"} Nov 24 07:25:00 crc kubenswrapper[4809]: I1124 07:25:00.907673 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 07:25:01 crc kubenswrapper[4809]: I1124 07:25:01.541152 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2" event={"ID":"a46047d8-9af2-4172-8e55-26dd29b587aa","Type":"ContainerStarted","Data":"1be18a96550209c8b66e0e407ec1ab86b71ab966a658966c3260c16cbe8a4394"} Nov 24 07:25:01 crc kubenswrapper[4809]: I1124 07:25:01.567628 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2" podStartSLOduration=2.171463121 podStartE2EDuration="2.567596573s" podCreationTimestamp="2025-11-24 07:24:59 +0000 UTC" firstStartedPulling="2025-11-24 07:25:00.506126879 +0000 UTC m=+1800.406718594" lastFinishedPulling="2025-11-24 07:25:00.902260351 +0000 UTC m=+1800.802852046" observedRunningTime="2025-11-24 07:25:01.561152682 +0000 UTC m=+1801.461744397" watchObservedRunningTime="2025-11-24 07:25:01.567596573 +0000 UTC m=+1801.468188318" Nov 24 07:25:09 crc kubenswrapper[4809]: I1124 07:25:09.818375 4809 scope.go:117] "RemoveContainer" containerID="212b977fdd911e9454a6815429587513ec64237aaa0ddea848608ae8ce5ca28d" Nov 24 07:25:11 crc kubenswrapper[4809]: I1124 07:25:11.651184 4809 generic.go:334] "Generic (PLEG): container finished" podID="a46047d8-9af2-4172-8e55-26dd29b587aa" containerID="1be18a96550209c8b66e0e407ec1ab86b71ab966a658966c3260c16cbe8a4394" exitCode=0 Nov 24 07:25:11 crc kubenswrapper[4809]: I1124 07:25:11.651315 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2" event={"ID":"a46047d8-9af2-4172-8e55-26dd29b587aa","Type":"ContainerDied","Data":"1be18a96550209c8b66e0e407ec1ab86b71ab966a658966c3260c16cbe8a4394"} Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.197231 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.320043 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a46047d8-9af2-4172-8e55-26dd29b587aa-inventory\") pod \"a46047d8-9af2-4172-8e55-26dd29b587aa\" (UID: \"a46047d8-9af2-4172-8e55-26dd29b587aa\") " Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.320087 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvhzg\" (UniqueName: \"kubernetes.io/projected/a46047d8-9af2-4172-8e55-26dd29b587aa-kube-api-access-dvhzg\") pod \"a46047d8-9af2-4172-8e55-26dd29b587aa\" (UID: \"a46047d8-9af2-4172-8e55-26dd29b587aa\") " Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.320185 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a46047d8-9af2-4172-8e55-26dd29b587aa-ssh-key\") pod \"a46047d8-9af2-4172-8e55-26dd29b587aa\" (UID: \"a46047d8-9af2-4172-8e55-26dd29b587aa\") " Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.330856 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a46047d8-9af2-4172-8e55-26dd29b587aa-kube-api-access-dvhzg" (OuterVolumeSpecName: "kube-api-access-dvhzg") pod "a46047d8-9af2-4172-8e55-26dd29b587aa" (UID: "a46047d8-9af2-4172-8e55-26dd29b587aa"). InnerVolumeSpecName "kube-api-access-dvhzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.354026 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a46047d8-9af2-4172-8e55-26dd29b587aa-inventory" (OuterVolumeSpecName: "inventory") pod "a46047d8-9af2-4172-8e55-26dd29b587aa" (UID: "a46047d8-9af2-4172-8e55-26dd29b587aa"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.356913 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a46047d8-9af2-4172-8e55-26dd29b587aa-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a46047d8-9af2-4172-8e55-26dd29b587aa" (UID: "a46047d8-9af2-4172-8e55-26dd29b587aa"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.422488 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a46047d8-9af2-4172-8e55-26dd29b587aa-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.422516 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvhzg\" (UniqueName: \"kubernetes.io/projected/a46047d8-9af2-4172-8e55-26dd29b587aa-kube-api-access-dvhzg\") on node \"crc\" DevicePath \"\"" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.422528 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a46047d8-9af2-4172-8e55-26dd29b587aa-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.678776 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2" event={"ID":"a46047d8-9af2-4172-8e55-26dd29b587aa","Type":"ContainerDied","Data":"1b86aa996fe6cd146225df4c1be8378d06f4d46fde1036d06fd4b41468a6e44d"} Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.678815 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b86aa996fe6cd146225df4c1be8378d06f4d46fde1036d06fd4b41468a6e44d" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.678903 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.758040 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9"] Nov 24 07:25:13 crc kubenswrapper[4809]: E1124 07:25:13.758502 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a46047d8-9af2-4172-8e55-26dd29b587aa" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.758524 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a46047d8-9af2-4172-8e55-26dd29b587aa" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.758709 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a46047d8-9af2-4172-8e55-26dd29b587aa" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.759366 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.761088 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.762223 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.764099 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lx7ll" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.764115 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.764238 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.764277 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.764590 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.764685 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.776992 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9"] Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.932072 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.932218 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.932348 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.932448 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.932507 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.932626 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvgq2\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-kube-api-access-cvgq2\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.932657 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.932702 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.932833 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.932896 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.932960 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.933127 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.933167 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:13 crc kubenswrapper[4809]: I1124 07:25:13.933264 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.034687 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.034779 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvgq2\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-kube-api-access-cvgq2\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.034807 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.034843 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.034911 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.034937 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.035002 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.035028 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.035051 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.035118 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.035163 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.035204 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.035238 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.035283 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.044583 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.044638 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.044819 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.045338 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.045463 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.045834 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.046406 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.046652 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.047028 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.047300 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.047805 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.048060 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.048263 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.056187 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvgq2\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-kube-api-access-cvgq2\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.080824 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.461085 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9"] Nov 24 07:25:14 crc kubenswrapper[4809]: I1124 07:25:14.688877 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" event={"ID":"d4e896d6-2d97-431e-bfaf-f48369c0af23","Type":"ContainerStarted","Data":"27de878d0b227708859e7a11abcc674b5aa61b34b389f2a7db22555987e99329"} Nov 24 07:25:15 crc kubenswrapper[4809]: I1124 07:25:15.699991 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" event={"ID":"d4e896d6-2d97-431e-bfaf-f48369c0af23","Type":"ContainerStarted","Data":"40a83a8d75e34ac77ed0cb564b15b68fc0088216072bcb5c501e9da323b1946b"} Nov 24 07:25:15 crc kubenswrapper[4809]: I1124 07:25:15.724621 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" podStartSLOduration=2.352651809 podStartE2EDuration="2.724604135s" podCreationTimestamp="2025-11-24 07:25:13 +0000 UTC" firstStartedPulling="2025-11-24 07:25:14.468245624 +0000 UTC m=+1814.368837329" lastFinishedPulling="2025-11-24 07:25:14.84019795 +0000 UTC m=+1814.740789655" observedRunningTime="2025-11-24 07:25:15.721685737 +0000 UTC m=+1815.622277442" watchObservedRunningTime="2025-11-24 07:25:15.724604135 +0000 UTC m=+1815.625195840" Nov 24 07:25:31 crc kubenswrapper[4809]: I1124 07:25:31.714159 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g2zmp"] Nov 24 07:25:31 crc kubenswrapper[4809]: I1124 07:25:31.717164 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g2zmp" Nov 24 07:25:31 crc kubenswrapper[4809]: I1124 07:25:31.734524 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g2zmp"] Nov 24 07:25:31 crc kubenswrapper[4809]: I1124 07:25:31.797004 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84021e2b-2492-410b-97e3-4f7a2532e61b-catalog-content\") pod \"redhat-operators-g2zmp\" (UID: \"84021e2b-2492-410b-97e3-4f7a2532e61b\") " pod="openshift-marketplace/redhat-operators-g2zmp" Nov 24 07:25:31 crc kubenswrapper[4809]: I1124 07:25:31.797086 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84021e2b-2492-410b-97e3-4f7a2532e61b-utilities\") pod \"redhat-operators-g2zmp\" (UID: \"84021e2b-2492-410b-97e3-4f7a2532e61b\") " pod="openshift-marketplace/redhat-operators-g2zmp" Nov 24 07:25:31 crc kubenswrapper[4809]: I1124 07:25:31.797142 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmrpk\" (UniqueName: \"kubernetes.io/projected/84021e2b-2492-410b-97e3-4f7a2532e61b-kube-api-access-xmrpk\") pod \"redhat-operators-g2zmp\" (UID: \"84021e2b-2492-410b-97e3-4f7a2532e61b\") " pod="openshift-marketplace/redhat-operators-g2zmp" Nov 24 07:25:31 crc kubenswrapper[4809]: I1124 07:25:31.898106 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84021e2b-2492-410b-97e3-4f7a2532e61b-catalog-content\") pod \"redhat-operators-g2zmp\" (UID: \"84021e2b-2492-410b-97e3-4f7a2532e61b\") " pod="openshift-marketplace/redhat-operators-g2zmp" Nov 24 07:25:31 crc kubenswrapper[4809]: I1124 07:25:31.898208 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84021e2b-2492-410b-97e3-4f7a2532e61b-utilities\") pod \"redhat-operators-g2zmp\" (UID: \"84021e2b-2492-410b-97e3-4f7a2532e61b\") " pod="openshift-marketplace/redhat-operators-g2zmp" Nov 24 07:25:31 crc kubenswrapper[4809]: I1124 07:25:31.898262 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmrpk\" (UniqueName: \"kubernetes.io/projected/84021e2b-2492-410b-97e3-4f7a2532e61b-kube-api-access-xmrpk\") pod \"redhat-operators-g2zmp\" (UID: \"84021e2b-2492-410b-97e3-4f7a2532e61b\") " pod="openshift-marketplace/redhat-operators-g2zmp" Nov 24 07:25:31 crc kubenswrapper[4809]: I1124 07:25:31.898811 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84021e2b-2492-410b-97e3-4f7a2532e61b-utilities\") pod \"redhat-operators-g2zmp\" (UID: \"84021e2b-2492-410b-97e3-4f7a2532e61b\") " pod="openshift-marketplace/redhat-operators-g2zmp" Nov 24 07:25:31 crc kubenswrapper[4809]: I1124 07:25:31.898905 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84021e2b-2492-410b-97e3-4f7a2532e61b-catalog-content\") pod \"redhat-operators-g2zmp\" (UID: \"84021e2b-2492-410b-97e3-4f7a2532e61b\") " pod="openshift-marketplace/redhat-operators-g2zmp" Nov 24 07:25:31 crc kubenswrapper[4809]: I1124 07:25:31.920861 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmrpk\" (UniqueName: \"kubernetes.io/projected/84021e2b-2492-410b-97e3-4f7a2532e61b-kube-api-access-xmrpk\") pod \"redhat-operators-g2zmp\" (UID: \"84021e2b-2492-410b-97e3-4f7a2532e61b\") " pod="openshift-marketplace/redhat-operators-g2zmp" Nov 24 07:25:32 crc kubenswrapper[4809]: I1124 07:25:32.066157 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g2zmp" Nov 24 07:25:32 crc kubenswrapper[4809]: I1124 07:25:32.511511 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g2zmp"] Nov 24 07:25:32 crc kubenswrapper[4809]: I1124 07:25:32.868526 4809 generic.go:334] "Generic (PLEG): container finished" podID="84021e2b-2492-410b-97e3-4f7a2532e61b" containerID="15aaa68f0b8894608134e84986de72718532e5b90f383dcb95f4a85749c0a9be" exitCode=0 Nov 24 07:25:32 crc kubenswrapper[4809]: I1124 07:25:32.868613 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g2zmp" event={"ID":"84021e2b-2492-410b-97e3-4f7a2532e61b","Type":"ContainerDied","Data":"15aaa68f0b8894608134e84986de72718532e5b90f383dcb95f4a85749c0a9be"} Nov 24 07:25:32 crc kubenswrapper[4809]: I1124 07:25:32.868821 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g2zmp" event={"ID":"84021e2b-2492-410b-97e3-4f7a2532e61b","Type":"ContainerStarted","Data":"f2c8192566d79dc498d31510a74017b96aa3f866f2782a088d180787c31f2412"} Nov 24 07:25:33 crc kubenswrapper[4809]: I1124 07:25:33.881358 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g2zmp" event={"ID":"84021e2b-2492-410b-97e3-4f7a2532e61b","Type":"ContainerStarted","Data":"4a8f9dd783036158ab532e1e3f41d32af90c81cae01e820b3302c6733e95fa61"} Nov 24 07:25:34 crc kubenswrapper[4809]: I1124 07:25:34.892311 4809 generic.go:334] "Generic (PLEG): container finished" podID="84021e2b-2492-410b-97e3-4f7a2532e61b" containerID="4a8f9dd783036158ab532e1e3f41d32af90c81cae01e820b3302c6733e95fa61" exitCode=0 Nov 24 07:25:34 crc kubenswrapper[4809]: I1124 07:25:34.910516 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g2zmp" event={"ID":"84021e2b-2492-410b-97e3-4f7a2532e61b","Type":"ContainerDied","Data":"4a8f9dd783036158ab532e1e3f41d32af90c81cae01e820b3302c6733e95fa61"} Nov 24 07:25:35 crc kubenswrapper[4809]: I1124 07:25:35.907011 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g2zmp" event={"ID":"84021e2b-2492-410b-97e3-4f7a2532e61b","Type":"ContainerStarted","Data":"8f8946a9718f15145833aa2f7679deea227c067da9a27ca6904a90b1c1e82591"} Nov 24 07:25:35 crc kubenswrapper[4809]: I1124 07:25:35.933555 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g2zmp" podStartSLOduration=2.464042611 podStartE2EDuration="4.933536467s" podCreationTimestamp="2025-11-24 07:25:31 +0000 UTC" firstStartedPulling="2025-11-24 07:25:32.871117592 +0000 UTC m=+1832.771709297" lastFinishedPulling="2025-11-24 07:25:35.340611408 +0000 UTC m=+1835.241203153" observedRunningTime="2025-11-24 07:25:35.92505017 +0000 UTC m=+1835.825641875" watchObservedRunningTime="2025-11-24 07:25:35.933536467 +0000 UTC m=+1835.834128172" Nov 24 07:25:42 crc kubenswrapper[4809]: I1124 07:25:42.067166 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g2zmp" Nov 24 07:25:42 crc kubenswrapper[4809]: I1124 07:25:42.067690 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g2zmp" Nov 24 07:25:42 crc kubenswrapper[4809]: I1124 07:25:42.140311 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g2zmp" Nov 24 07:25:43 crc kubenswrapper[4809]: I1124 07:25:43.018216 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g2zmp" Nov 24 07:25:43 crc kubenswrapper[4809]: I1124 07:25:43.076275 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g2zmp"] Nov 24 07:25:44 crc kubenswrapper[4809]: I1124 07:25:44.977561 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-g2zmp" podUID="84021e2b-2492-410b-97e3-4f7a2532e61b" containerName="registry-server" containerID="cri-o://8f8946a9718f15145833aa2f7679deea227c067da9a27ca6904a90b1c1e82591" gracePeriod=2 Nov 24 07:25:45 crc kubenswrapper[4809]: I1124 07:25:45.464117 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g2zmp" Nov 24 07:25:45 crc kubenswrapper[4809]: I1124 07:25:45.585236 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84021e2b-2492-410b-97e3-4f7a2532e61b-catalog-content\") pod \"84021e2b-2492-410b-97e3-4f7a2532e61b\" (UID: \"84021e2b-2492-410b-97e3-4f7a2532e61b\") " Nov 24 07:25:45 crc kubenswrapper[4809]: I1124 07:25:45.585833 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmrpk\" (UniqueName: \"kubernetes.io/projected/84021e2b-2492-410b-97e3-4f7a2532e61b-kube-api-access-xmrpk\") pod \"84021e2b-2492-410b-97e3-4f7a2532e61b\" (UID: \"84021e2b-2492-410b-97e3-4f7a2532e61b\") " Nov 24 07:25:45 crc kubenswrapper[4809]: I1124 07:25:45.586040 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84021e2b-2492-410b-97e3-4f7a2532e61b-utilities\") pod \"84021e2b-2492-410b-97e3-4f7a2532e61b\" (UID: \"84021e2b-2492-410b-97e3-4f7a2532e61b\") " Nov 24 07:25:45 crc kubenswrapper[4809]: I1124 07:25:45.586606 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84021e2b-2492-410b-97e3-4f7a2532e61b-utilities" (OuterVolumeSpecName: "utilities") pod "84021e2b-2492-410b-97e3-4f7a2532e61b" (UID: "84021e2b-2492-410b-97e3-4f7a2532e61b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:25:45 crc kubenswrapper[4809]: I1124 07:25:45.587148 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84021e2b-2492-410b-97e3-4f7a2532e61b-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:25:45 crc kubenswrapper[4809]: I1124 07:25:45.591004 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84021e2b-2492-410b-97e3-4f7a2532e61b-kube-api-access-xmrpk" (OuterVolumeSpecName: "kube-api-access-xmrpk") pod "84021e2b-2492-410b-97e3-4f7a2532e61b" (UID: "84021e2b-2492-410b-97e3-4f7a2532e61b"). InnerVolumeSpecName "kube-api-access-xmrpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:25:45 crc kubenswrapper[4809]: I1124 07:25:45.673760 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84021e2b-2492-410b-97e3-4f7a2532e61b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "84021e2b-2492-410b-97e3-4f7a2532e61b" (UID: "84021e2b-2492-410b-97e3-4f7a2532e61b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:25:45 crc kubenswrapper[4809]: I1124 07:25:45.689296 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmrpk\" (UniqueName: \"kubernetes.io/projected/84021e2b-2492-410b-97e3-4f7a2532e61b-kube-api-access-xmrpk\") on node \"crc\" DevicePath \"\"" Nov 24 07:25:45 crc kubenswrapper[4809]: I1124 07:25:45.689326 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84021e2b-2492-410b-97e3-4f7a2532e61b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:25:45 crc kubenswrapper[4809]: I1124 07:25:45.987718 4809 generic.go:334] "Generic (PLEG): container finished" podID="84021e2b-2492-410b-97e3-4f7a2532e61b" containerID="8f8946a9718f15145833aa2f7679deea227c067da9a27ca6904a90b1c1e82591" exitCode=0 Nov 24 07:25:45 crc kubenswrapper[4809]: I1124 07:25:45.987768 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g2zmp" event={"ID":"84021e2b-2492-410b-97e3-4f7a2532e61b","Type":"ContainerDied","Data":"8f8946a9718f15145833aa2f7679deea227c067da9a27ca6904a90b1c1e82591"} Nov 24 07:25:45 crc kubenswrapper[4809]: I1124 07:25:45.987807 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g2zmp" event={"ID":"84021e2b-2492-410b-97e3-4f7a2532e61b","Type":"ContainerDied","Data":"f2c8192566d79dc498d31510a74017b96aa3f866f2782a088d180787c31f2412"} Nov 24 07:25:45 crc kubenswrapper[4809]: I1124 07:25:45.987832 4809 scope.go:117] "RemoveContainer" containerID="8f8946a9718f15145833aa2f7679deea227c067da9a27ca6904a90b1c1e82591" Nov 24 07:25:45 crc kubenswrapper[4809]: I1124 07:25:45.988141 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g2zmp" Nov 24 07:25:46 crc kubenswrapper[4809]: I1124 07:25:46.010091 4809 scope.go:117] "RemoveContainer" containerID="4a8f9dd783036158ab532e1e3f41d32af90c81cae01e820b3302c6733e95fa61" Nov 24 07:25:46 crc kubenswrapper[4809]: I1124 07:25:46.031485 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g2zmp"] Nov 24 07:25:46 crc kubenswrapper[4809]: I1124 07:25:46.041326 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-g2zmp"] Nov 24 07:25:46 crc kubenswrapper[4809]: I1124 07:25:46.047213 4809 scope.go:117] "RemoveContainer" containerID="15aaa68f0b8894608134e84986de72718532e5b90f383dcb95f4a85749c0a9be" Nov 24 07:25:46 crc kubenswrapper[4809]: I1124 07:25:46.101028 4809 scope.go:117] "RemoveContainer" containerID="8f8946a9718f15145833aa2f7679deea227c067da9a27ca6904a90b1c1e82591" Nov 24 07:25:46 crc kubenswrapper[4809]: E1124 07:25:46.101480 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f8946a9718f15145833aa2f7679deea227c067da9a27ca6904a90b1c1e82591\": container with ID starting with 8f8946a9718f15145833aa2f7679deea227c067da9a27ca6904a90b1c1e82591 not found: ID does not exist" containerID="8f8946a9718f15145833aa2f7679deea227c067da9a27ca6904a90b1c1e82591" Nov 24 07:25:46 crc kubenswrapper[4809]: I1124 07:25:46.101533 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f8946a9718f15145833aa2f7679deea227c067da9a27ca6904a90b1c1e82591"} err="failed to get container status \"8f8946a9718f15145833aa2f7679deea227c067da9a27ca6904a90b1c1e82591\": rpc error: code = NotFound desc = could not find container \"8f8946a9718f15145833aa2f7679deea227c067da9a27ca6904a90b1c1e82591\": container with ID starting with 8f8946a9718f15145833aa2f7679deea227c067da9a27ca6904a90b1c1e82591 not found: ID does not exist" Nov 24 07:25:46 crc kubenswrapper[4809]: I1124 07:25:46.101597 4809 scope.go:117] "RemoveContainer" containerID="4a8f9dd783036158ab532e1e3f41d32af90c81cae01e820b3302c6733e95fa61" Nov 24 07:25:46 crc kubenswrapper[4809]: E1124 07:25:46.102030 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a8f9dd783036158ab532e1e3f41d32af90c81cae01e820b3302c6733e95fa61\": container with ID starting with 4a8f9dd783036158ab532e1e3f41d32af90c81cae01e820b3302c6733e95fa61 not found: ID does not exist" containerID="4a8f9dd783036158ab532e1e3f41d32af90c81cae01e820b3302c6733e95fa61" Nov 24 07:25:46 crc kubenswrapper[4809]: I1124 07:25:46.102061 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a8f9dd783036158ab532e1e3f41d32af90c81cae01e820b3302c6733e95fa61"} err="failed to get container status \"4a8f9dd783036158ab532e1e3f41d32af90c81cae01e820b3302c6733e95fa61\": rpc error: code = NotFound desc = could not find container \"4a8f9dd783036158ab532e1e3f41d32af90c81cae01e820b3302c6733e95fa61\": container with ID starting with 4a8f9dd783036158ab532e1e3f41d32af90c81cae01e820b3302c6733e95fa61 not found: ID does not exist" Nov 24 07:25:46 crc kubenswrapper[4809]: I1124 07:25:46.102084 4809 scope.go:117] "RemoveContainer" containerID="15aaa68f0b8894608134e84986de72718532e5b90f383dcb95f4a85749c0a9be" Nov 24 07:25:46 crc kubenswrapper[4809]: E1124 07:25:46.102573 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15aaa68f0b8894608134e84986de72718532e5b90f383dcb95f4a85749c0a9be\": container with ID starting with 15aaa68f0b8894608134e84986de72718532e5b90f383dcb95f4a85749c0a9be not found: ID does not exist" containerID="15aaa68f0b8894608134e84986de72718532e5b90f383dcb95f4a85749c0a9be" Nov 24 07:25:46 crc kubenswrapper[4809]: I1124 07:25:46.102728 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15aaa68f0b8894608134e84986de72718532e5b90f383dcb95f4a85749c0a9be"} err="failed to get container status \"15aaa68f0b8894608134e84986de72718532e5b90f383dcb95f4a85749c0a9be\": rpc error: code = NotFound desc = could not find container \"15aaa68f0b8894608134e84986de72718532e5b90f383dcb95f4a85749c0a9be\": container with ID starting with 15aaa68f0b8894608134e84986de72718532e5b90f383dcb95f4a85749c0a9be not found: ID does not exist" Nov 24 07:25:46 crc kubenswrapper[4809]: I1124 07:25:46.900398 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84021e2b-2492-410b-97e3-4f7a2532e61b" path="/var/lib/kubelet/pods/84021e2b-2492-410b-97e3-4f7a2532e61b/volumes" Nov 24 07:26:00 crc kubenswrapper[4809]: I1124 07:26:00.134671 4809 generic.go:334] "Generic (PLEG): container finished" podID="d4e896d6-2d97-431e-bfaf-f48369c0af23" containerID="40a83a8d75e34ac77ed0cb564b15b68fc0088216072bcb5c501e9da323b1946b" exitCode=0 Nov 24 07:26:00 crc kubenswrapper[4809]: I1124 07:26:00.134704 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" event={"ID":"d4e896d6-2d97-431e-bfaf-f48369c0af23","Type":"ContainerDied","Data":"40a83a8d75e34ac77ed0cb564b15b68fc0088216072bcb5c501e9da323b1946b"} Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.639381 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.813686 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-nova-combined-ca-bundle\") pod \"d4e896d6-2d97-431e-bfaf-f48369c0af23\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.814171 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"d4e896d6-2d97-431e-bfaf-f48369c0af23\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.814519 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-ssh-key\") pod \"d4e896d6-2d97-431e-bfaf-f48369c0af23\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.814664 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-repo-setup-combined-ca-bundle\") pod \"d4e896d6-2d97-431e-bfaf-f48369c0af23\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.814803 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-ovn-combined-ca-bundle\") pod \"d4e896d6-2d97-431e-bfaf-f48369c0af23\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.814940 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"d4e896d6-2d97-431e-bfaf-f48369c0af23\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.815064 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-libvirt-combined-ca-bundle\") pod \"d4e896d6-2d97-431e-bfaf-f48369c0af23\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.815217 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-ovn-default-certs-0\") pod \"d4e896d6-2d97-431e-bfaf-f48369c0af23\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.815384 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-inventory\") pod \"d4e896d6-2d97-431e-bfaf-f48369c0af23\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.815538 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvgq2\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-kube-api-access-cvgq2\") pod \"d4e896d6-2d97-431e-bfaf-f48369c0af23\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.815652 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-neutron-metadata-combined-ca-bundle\") pod \"d4e896d6-2d97-431e-bfaf-f48369c0af23\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.815783 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-telemetry-combined-ca-bundle\") pod \"d4e896d6-2d97-431e-bfaf-f48369c0af23\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.815907 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"d4e896d6-2d97-431e-bfaf-f48369c0af23\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.816037 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-bootstrap-combined-ca-bundle\") pod \"d4e896d6-2d97-431e-bfaf-f48369c0af23\" (UID: \"d4e896d6-2d97-431e-bfaf-f48369c0af23\") " Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.820843 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "d4e896d6-2d97-431e-bfaf-f48369c0af23" (UID: "d4e896d6-2d97-431e-bfaf-f48369c0af23"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.822134 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "d4e896d6-2d97-431e-bfaf-f48369c0af23" (UID: "d4e896d6-2d97-431e-bfaf-f48369c0af23"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.822182 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "d4e896d6-2d97-431e-bfaf-f48369c0af23" (UID: "d4e896d6-2d97-431e-bfaf-f48369c0af23"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.822643 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "d4e896d6-2d97-431e-bfaf-f48369c0af23" (UID: "d4e896d6-2d97-431e-bfaf-f48369c0af23"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.822680 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "d4e896d6-2d97-431e-bfaf-f48369c0af23" (UID: "d4e896d6-2d97-431e-bfaf-f48369c0af23"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.822697 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "d4e896d6-2d97-431e-bfaf-f48369c0af23" (UID: "d4e896d6-2d97-431e-bfaf-f48369c0af23"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.823494 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-kube-api-access-cvgq2" (OuterVolumeSpecName: "kube-api-access-cvgq2") pod "d4e896d6-2d97-431e-bfaf-f48369c0af23" (UID: "d4e896d6-2d97-431e-bfaf-f48369c0af23"). InnerVolumeSpecName "kube-api-access-cvgq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.823821 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "d4e896d6-2d97-431e-bfaf-f48369c0af23" (UID: "d4e896d6-2d97-431e-bfaf-f48369c0af23"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.824413 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "d4e896d6-2d97-431e-bfaf-f48369c0af23" (UID: "d4e896d6-2d97-431e-bfaf-f48369c0af23"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.824466 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "d4e896d6-2d97-431e-bfaf-f48369c0af23" (UID: "d4e896d6-2d97-431e-bfaf-f48369c0af23"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.825767 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "d4e896d6-2d97-431e-bfaf-f48369c0af23" (UID: "d4e896d6-2d97-431e-bfaf-f48369c0af23"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.830421 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "d4e896d6-2d97-431e-bfaf-f48369c0af23" (UID: "d4e896d6-2d97-431e-bfaf-f48369c0af23"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.844677 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d4e896d6-2d97-431e-bfaf-f48369c0af23" (UID: "d4e896d6-2d97-431e-bfaf-f48369c0af23"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.863212 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-inventory" (OuterVolumeSpecName: "inventory") pod "d4e896d6-2d97-431e-bfaf-f48369c0af23" (UID: "d4e896d6-2d97-431e-bfaf-f48369c0af23"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.919555 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.919594 4809 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.919609 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.919624 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.919636 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvgq2\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-kube-api-access-cvgq2\") on node \"crc\" DevicePath \"\"" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.919649 4809 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.919661 4809 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.919673 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.919687 4809 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.919700 4809 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.919711 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d4e896d6-2d97-431e-bfaf-f48369c0af23-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.919725 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.919739 4809 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:26:01 crc kubenswrapper[4809]: I1124 07:26:01.919753 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e896d6-2d97-431e-bfaf-f48369c0af23-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.163586 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" event={"ID":"d4e896d6-2d97-431e-bfaf-f48369c0af23","Type":"ContainerDied","Data":"27de878d0b227708859e7a11abcc674b5aa61b34b389f2a7db22555987e99329"} Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.163643 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27de878d0b227708859e7a11abcc674b5aa61b34b389f2a7db22555987e99329" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.163722 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.282027 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm"] Nov 24 07:26:02 crc kubenswrapper[4809]: E1124 07:26:02.282506 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4e896d6-2d97-431e-bfaf-f48369c0af23" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.282531 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4e896d6-2d97-431e-bfaf-f48369c0af23" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 07:26:02 crc kubenswrapper[4809]: E1124 07:26:02.282553 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84021e2b-2492-410b-97e3-4f7a2532e61b" containerName="extract-utilities" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.282563 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="84021e2b-2492-410b-97e3-4f7a2532e61b" containerName="extract-utilities" Nov 24 07:26:02 crc kubenswrapper[4809]: E1124 07:26:02.282596 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84021e2b-2492-410b-97e3-4f7a2532e61b" containerName="extract-content" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.282604 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="84021e2b-2492-410b-97e3-4f7a2532e61b" containerName="extract-content" Nov 24 07:26:02 crc kubenswrapper[4809]: E1124 07:26:02.282624 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84021e2b-2492-410b-97e3-4f7a2532e61b" containerName="registry-server" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.282631 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="84021e2b-2492-410b-97e3-4f7a2532e61b" containerName="registry-server" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.282835 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="84021e2b-2492-410b-97e3-4f7a2532e61b" containerName="registry-server" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.282853 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4e896d6-2d97-431e-bfaf-f48369c0af23" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.283695 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.286610 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.286818 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.286987 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.287183 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.287381 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lx7ll" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.303303 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm"] Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.435495 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mfjkm\" (UID: \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.435594 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mfjkm\" (UID: \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.435655 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mdlk\" (UniqueName: \"kubernetes.io/projected/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-kube-api-access-6mdlk\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mfjkm\" (UID: \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.435723 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mfjkm\" (UID: \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.435783 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mfjkm\" (UID: \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.536887 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mdlk\" (UniqueName: \"kubernetes.io/projected/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-kube-api-access-6mdlk\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mfjkm\" (UID: \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.536999 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mfjkm\" (UID: \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.537076 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mfjkm\" (UID: \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.537138 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mfjkm\" (UID: \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.537215 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mfjkm\" (UID: \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.537810 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mfjkm\" (UID: \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.542740 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mfjkm\" (UID: \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.544475 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mfjkm\" (UID: \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.544958 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mfjkm\" (UID: \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.558254 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mdlk\" (UniqueName: \"kubernetes.io/projected/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-kube-api-access-6mdlk\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mfjkm\" (UID: \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" Nov 24 07:26:02 crc kubenswrapper[4809]: I1124 07:26:02.607277 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" Nov 24 07:26:03 crc kubenswrapper[4809]: I1124 07:26:03.187913 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm"] Nov 24 07:26:04 crc kubenswrapper[4809]: I1124 07:26:04.182463 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" event={"ID":"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255","Type":"ContainerStarted","Data":"16d7de3eead0ba40ede812c4fc90d1140bcda58afaf54f5c1b5d1952e6e080cf"} Nov 24 07:26:04 crc kubenswrapper[4809]: I1124 07:26:04.182763 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" event={"ID":"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255","Type":"ContainerStarted","Data":"4df16d775219f6b5dc8f800f0ff0d1394903fa6eb46dc595d50f901ea0b44124"} Nov 24 07:26:04 crc kubenswrapper[4809]: I1124 07:26:04.211235 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" podStartSLOduration=1.742638999 podStartE2EDuration="2.211203076s" podCreationTimestamp="2025-11-24 07:26:02 +0000 UTC" firstStartedPulling="2025-11-24 07:26:03.193574252 +0000 UTC m=+1863.094165957" lastFinishedPulling="2025-11-24 07:26:03.662138319 +0000 UTC m=+1863.562730034" observedRunningTime="2025-11-24 07:26:04.198019223 +0000 UTC m=+1864.098610938" watchObservedRunningTime="2025-11-24 07:26:04.211203076 +0000 UTC m=+1864.111794821" Nov 24 07:27:15 crc kubenswrapper[4809]: I1124 07:27:15.843779 4809 generic.go:334] "Generic (PLEG): container finished" podID="a25b9532-a2c1-4e47-b8c2-7c5cec0fa255" containerID="16d7de3eead0ba40ede812c4fc90d1140bcda58afaf54f5c1b5d1952e6e080cf" exitCode=0 Nov 24 07:27:15 crc kubenswrapper[4809]: I1124 07:27:15.843821 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" event={"ID":"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255","Type":"ContainerDied","Data":"16d7de3eead0ba40ede812c4fc90d1140bcda58afaf54f5c1b5d1952e6e080cf"} Nov 24 07:27:17 crc kubenswrapper[4809]: I1124 07:27:17.269293 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" Nov 24 07:27:17 crc kubenswrapper[4809]: I1124 07:27:17.407329 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-inventory\") pod \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\" (UID: \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\") " Nov 24 07:27:17 crc kubenswrapper[4809]: I1124 07:27:17.407512 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-ovncontroller-config-0\") pod \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\" (UID: \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\") " Nov 24 07:27:17 crc kubenswrapper[4809]: I1124 07:27:17.407567 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-ssh-key\") pod \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\" (UID: \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\") " Nov 24 07:27:17 crc kubenswrapper[4809]: I1124 07:27:17.407651 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-ovn-combined-ca-bundle\") pod \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\" (UID: \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\") " Nov 24 07:27:17 crc kubenswrapper[4809]: I1124 07:27:17.407771 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mdlk\" (UniqueName: \"kubernetes.io/projected/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-kube-api-access-6mdlk\") pod \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\" (UID: \"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255\") " Nov 24 07:27:17 crc kubenswrapper[4809]: I1124 07:27:17.413398 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-kube-api-access-6mdlk" (OuterVolumeSpecName: "kube-api-access-6mdlk") pod "a25b9532-a2c1-4e47-b8c2-7c5cec0fa255" (UID: "a25b9532-a2c1-4e47-b8c2-7c5cec0fa255"). InnerVolumeSpecName "kube-api-access-6mdlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:27:17 crc kubenswrapper[4809]: I1124 07:27:17.413672 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "a25b9532-a2c1-4e47-b8c2-7c5cec0fa255" (UID: "a25b9532-a2c1-4e47-b8c2-7c5cec0fa255"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:27:17 crc kubenswrapper[4809]: I1124 07:27:17.436208 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-inventory" (OuterVolumeSpecName: "inventory") pod "a25b9532-a2c1-4e47-b8c2-7c5cec0fa255" (UID: "a25b9532-a2c1-4e47-b8c2-7c5cec0fa255"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:27:17 crc kubenswrapper[4809]: I1124 07:27:17.441259 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "a25b9532-a2c1-4e47-b8c2-7c5cec0fa255" (UID: "a25b9532-a2c1-4e47-b8c2-7c5cec0fa255"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:27:17 crc kubenswrapper[4809]: I1124 07:27:17.463254 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a25b9532-a2c1-4e47-b8c2-7c5cec0fa255" (UID: "a25b9532-a2c1-4e47-b8c2-7c5cec0fa255"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:27:17 crc kubenswrapper[4809]: I1124 07:27:17.511011 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mdlk\" (UniqueName: \"kubernetes.io/projected/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-kube-api-access-6mdlk\") on node \"crc\" DevicePath \"\"" Nov 24 07:27:17 crc kubenswrapper[4809]: I1124 07:27:17.511094 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 07:27:17 crc kubenswrapper[4809]: I1124 07:27:17.511114 4809 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:27:17 crc kubenswrapper[4809]: I1124 07:27:17.511129 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 07:27:17 crc kubenswrapper[4809]: I1124 07:27:17.511148 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a25b9532-a2c1-4e47-b8c2-7c5cec0fa255-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:27:17 crc kubenswrapper[4809]: I1124 07:27:17.866806 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" event={"ID":"a25b9532-a2c1-4e47-b8c2-7c5cec0fa255","Type":"ContainerDied","Data":"4df16d775219f6b5dc8f800f0ff0d1394903fa6eb46dc595d50f901ea0b44124"} Nov 24 07:27:17 crc kubenswrapper[4809]: I1124 07:27:17.866847 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4df16d775219f6b5dc8f800f0ff0d1394903fa6eb46dc595d50f901ea0b44124" Nov 24 07:27:17 crc kubenswrapper[4809]: I1124 07:27:17.866905 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mfjkm" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.042226 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x"] Nov 24 07:27:18 crc kubenswrapper[4809]: E1124 07:27:18.042668 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a25b9532-a2c1-4e47-b8c2-7c5cec0fa255" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.042689 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a25b9532-a2c1-4e47-b8c2-7c5cec0fa255" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.042958 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a25b9532-a2c1-4e47-b8c2-7c5cec0fa255" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.043616 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.043674 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.043892 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.051558 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.053054 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.053815 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lx7ll" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.054236 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.054262 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.054383 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.074381 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x"] Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.123298 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.123493 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78g5j\" (UniqueName: \"kubernetes.io/projected/aef096ab-d8a4-4b78-97be-17967897eb98-kube-api-access-78g5j\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.123548 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.123592 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.123654 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.123743 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.225600 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.225694 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.225763 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78g5j\" (UniqueName: \"kubernetes.io/projected/aef096ab-d8a4-4b78-97be-17967897eb98-kube-api-access-78g5j\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.225790 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.225810 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.225879 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.230666 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.231356 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.231471 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.232010 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.234805 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.243764 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78g5j\" (UniqueName: \"kubernetes.io/projected/aef096ab-d8a4-4b78-97be-17967897eb98-kube-api-access-78g5j\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.364996 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.887575 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x"] Nov 24 07:27:18 crc kubenswrapper[4809]: I1124 07:27:18.898356 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 07:27:19 crc kubenswrapper[4809]: I1124 07:27:19.886099 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" event={"ID":"aef096ab-d8a4-4b78-97be-17967897eb98","Type":"ContainerStarted","Data":"0c8274c6a3541b140b8b724880423ba72cb640e5d2fc4e0c01c2ba37a1fadf52"} Nov 24 07:27:19 crc kubenswrapper[4809]: I1124 07:27:19.887864 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" event={"ID":"aef096ab-d8a4-4b78-97be-17967897eb98","Type":"ContainerStarted","Data":"c6c70130e055e4f3b2793e18e214f1a40d3378b6815ecf47588adb1bb42b37ea"} Nov 24 07:27:19 crc kubenswrapper[4809]: I1124 07:27:19.904837 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" podStartSLOduration=1.404636571 podStartE2EDuration="1.904818093s" podCreationTimestamp="2025-11-24 07:27:18 +0000 UTC" firstStartedPulling="2025-11-24 07:27:18.897554456 +0000 UTC m=+1938.798146211" lastFinishedPulling="2025-11-24 07:27:19.397736028 +0000 UTC m=+1939.298327733" observedRunningTime="2025-11-24 07:27:19.902284435 +0000 UTC m=+1939.802876180" watchObservedRunningTime="2025-11-24 07:27:19.904818093 +0000 UTC m=+1939.805409788" Nov 24 07:27:48 crc kubenswrapper[4809]: I1124 07:27:48.043581 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:27:48 crc kubenswrapper[4809]: I1124 07:27:48.044090 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:28:14 crc kubenswrapper[4809]: E1124 07:28:14.598337 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaef096ab_d8a4_4b78_97be_17967897eb98.slice/crio-0c8274c6a3541b140b8b724880423ba72cb640e5d2fc4e0c01c2ba37a1fadf52.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaef096ab_d8a4_4b78_97be_17967897eb98.slice/crio-conmon-0c8274c6a3541b140b8b724880423ba72cb640e5d2fc4e0c01c2ba37a1fadf52.scope\": RecentStats: unable to find data in memory cache]" Nov 24 07:28:15 crc kubenswrapper[4809]: I1124 07:28:15.390804 4809 generic.go:334] "Generic (PLEG): container finished" podID="aef096ab-d8a4-4b78-97be-17967897eb98" containerID="0c8274c6a3541b140b8b724880423ba72cb640e5d2fc4e0c01c2ba37a1fadf52" exitCode=0 Nov 24 07:28:15 crc kubenswrapper[4809]: I1124 07:28:15.390926 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" event={"ID":"aef096ab-d8a4-4b78-97be-17967897eb98","Type":"ContainerDied","Data":"0c8274c6a3541b140b8b724880423ba72cb640e5d2fc4e0c01c2ba37a1fadf52"} Nov 24 07:28:16 crc kubenswrapper[4809]: I1124 07:28:16.875177 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" Nov 24 07:28:16 crc kubenswrapper[4809]: I1124 07:28:16.970728 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-nova-metadata-neutron-config-0\") pod \"aef096ab-d8a4-4b78-97be-17967897eb98\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " Nov 24 07:28:16 crc kubenswrapper[4809]: I1124 07:28:16.970792 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78g5j\" (UniqueName: \"kubernetes.io/projected/aef096ab-d8a4-4b78-97be-17967897eb98-kube-api-access-78g5j\") pod \"aef096ab-d8a4-4b78-97be-17967897eb98\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " Nov 24 07:28:16 crc kubenswrapper[4809]: I1124 07:28:16.970883 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-neutron-ovn-metadata-agent-neutron-config-0\") pod \"aef096ab-d8a4-4b78-97be-17967897eb98\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " Nov 24 07:28:16 crc kubenswrapper[4809]: I1124 07:28:16.971100 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-ssh-key\") pod \"aef096ab-d8a4-4b78-97be-17967897eb98\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " Nov 24 07:28:16 crc kubenswrapper[4809]: I1124 07:28:16.971754 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-inventory\") pod \"aef096ab-d8a4-4b78-97be-17967897eb98\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " Nov 24 07:28:16 crc kubenswrapper[4809]: I1124 07:28:16.971855 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-neutron-metadata-combined-ca-bundle\") pod \"aef096ab-d8a4-4b78-97be-17967897eb98\" (UID: \"aef096ab-d8a4-4b78-97be-17967897eb98\") " Nov 24 07:28:16 crc kubenswrapper[4809]: I1124 07:28:16.976951 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "aef096ab-d8a4-4b78-97be-17967897eb98" (UID: "aef096ab-d8a4-4b78-97be-17967897eb98"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:28:16 crc kubenswrapper[4809]: I1124 07:28:16.977648 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aef096ab-d8a4-4b78-97be-17967897eb98-kube-api-access-78g5j" (OuterVolumeSpecName: "kube-api-access-78g5j") pod "aef096ab-d8a4-4b78-97be-17967897eb98" (UID: "aef096ab-d8a4-4b78-97be-17967897eb98"). InnerVolumeSpecName "kube-api-access-78g5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.004567 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-inventory" (OuterVolumeSpecName: "inventory") pod "aef096ab-d8a4-4b78-97be-17967897eb98" (UID: "aef096ab-d8a4-4b78-97be-17967897eb98"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.005038 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "aef096ab-d8a4-4b78-97be-17967897eb98" (UID: "aef096ab-d8a4-4b78-97be-17967897eb98"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.015544 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "aef096ab-d8a4-4b78-97be-17967897eb98" (UID: "aef096ab-d8a4-4b78-97be-17967897eb98"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.025945 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "aef096ab-d8a4-4b78-97be-17967897eb98" (UID: "aef096ab-d8a4-4b78-97be-17967897eb98"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.080282 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.080331 4809 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.080349 4809 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.080388 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78g5j\" (UniqueName: \"kubernetes.io/projected/aef096ab-d8a4-4b78-97be-17967897eb98-kube-api-access-78g5j\") on node \"crc\" DevicePath \"\"" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.080404 4809 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.080417 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aef096ab-d8a4-4b78-97be-17967897eb98-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.408871 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" event={"ID":"aef096ab-d8a4-4b78-97be-17967897eb98","Type":"ContainerDied","Data":"c6c70130e055e4f3b2793e18e214f1a40d3378b6815ecf47588adb1bb42b37ea"} Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.408923 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6c70130e055e4f3b2793e18e214f1a40d3378b6815ecf47588adb1bb42b37ea" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.409071 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.503023 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s"] Nov 24 07:28:17 crc kubenswrapper[4809]: E1124 07:28:17.503421 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aef096ab-d8a4-4b78-97be-17967897eb98" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.503442 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="aef096ab-d8a4-4b78-97be-17967897eb98" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.503683 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="aef096ab-d8a4-4b78-97be-17967897eb98" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.504485 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.507521 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.507525 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.507738 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.507940 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lx7ll" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.508178 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.523809 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s"] Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.598489 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s\" (UID: \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.598564 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s\" (UID: \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.598705 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s\" (UID: \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.598775 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s\" (UID: \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.599091 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggxb4\" (UniqueName: \"kubernetes.io/projected/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-kube-api-access-ggxb4\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s\" (UID: \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.700534 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s\" (UID: \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.700596 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s\" (UID: \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.700630 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s\" (UID: \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.700741 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggxb4\" (UniqueName: \"kubernetes.io/projected/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-kube-api-access-ggxb4\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s\" (UID: \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.700824 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s\" (UID: \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.704773 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s\" (UID: \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.705681 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s\" (UID: \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.706833 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s\" (UID: \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.708246 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s\" (UID: \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.722123 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggxb4\" (UniqueName: \"kubernetes.io/projected/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-kube-api-access-ggxb4\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s\" (UID: \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" Nov 24 07:28:17 crc kubenswrapper[4809]: I1124 07:28:17.827662 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" Nov 24 07:28:18 crc kubenswrapper[4809]: I1124 07:28:18.043148 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:28:18 crc kubenswrapper[4809]: I1124 07:28:18.043213 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:28:18 crc kubenswrapper[4809]: I1124 07:28:18.043261 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 07:28:18 crc kubenswrapper[4809]: I1124 07:28:18.044070 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"da763efb03a569decc0792fe26fc2e80b6b2ae304146cd9bddaf55c16474daac"} pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:28:18 crc kubenswrapper[4809]: I1124 07:28:18.044153 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" containerID="cri-o://da763efb03a569decc0792fe26fc2e80b6b2ae304146cd9bddaf55c16474daac" gracePeriod=600 Nov 24 07:28:18 crc kubenswrapper[4809]: I1124 07:28:18.316406 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s"] Nov 24 07:28:18 crc kubenswrapper[4809]: W1124 07:28:18.317635 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd55a20b_bdd2_40bf_85d7_c0b485352b8f.slice/crio-5836f266843a9c18248bacd679bf75921ccbdd00f3b48a8df366ea542f6be54b WatchSource:0}: Error finding container 5836f266843a9c18248bacd679bf75921ccbdd00f3b48a8df366ea542f6be54b: Status 404 returned error can't find the container with id 5836f266843a9c18248bacd679bf75921ccbdd00f3b48a8df366ea542f6be54b Nov 24 07:28:18 crc kubenswrapper[4809]: I1124 07:28:18.419251 4809 generic.go:334] "Generic (PLEG): container finished" podID="be8050d5-1aea-453e-a06e-87702b0e856b" containerID="da763efb03a569decc0792fe26fc2e80b6b2ae304146cd9bddaf55c16474daac" exitCode=0 Nov 24 07:28:18 crc kubenswrapper[4809]: I1124 07:28:18.419325 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerDied","Data":"da763efb03a569decc0792fe26fc2e80b6b2ae304146cd9bddaf55c16474daac"} Nov 24 07:28:18 crc kubenswrapper[4809]: I1124 07:28:18.419563 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerStarted","Data":"b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add"} Nov 24 07:28:18 crc kubenswrapper[4809]: I1124 07:28:18.419634 4809 scope.go:117] "RemoveContainer" containerID="e5e33fa1e11f695e1caa6ba8ddb9fa5aac30479a6651dd4d423451b997045b6e" Nov 24 07:28:18 crc kubenswrapper[4809]: I1124 07:28:18.423501 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" event={"ID":"fd55a20b-bdd2-40bf-85d7-c0b485352b8f","Type":"ContainerStarted","Data":"5836f266843a9c18248bacd679bf75921ccbdd00f3b48a8df366ea542f6be54b"} Nov 24 07:28:19 crc kubenswrapper[4809]: I1124 07:28:19.434067 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" event={"ID":"fd55a20b-bdd2-40bf-85d7-c0b485352b8f","Type":"ContainerStarted","Data":"83dd7678b8b5ef6575cc667a269d20ce80314a351a65815e6eb46f83d56c9e3f"} Nov 24 07:28:19 crc kubenswrapper[4809]: I1124 07:28:19.451665 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" podStartSLOduration=1.9356371559999999 podStartE2EDuration="2.45164683s" podCreationTimestamp="2025-11-24 07:28:17 +0000 UTC" firstStartedPulling="2025-11-24 07:28:18.319820266 +0000 UTC m=+1998.220411991" lastFinishedPulling="2025-11-24 07:28:18.83582989 +0000 UTC m=+1998.736421665" observedRunningTime="2025-11-24 07:28:19.448137146 +0000 UTC m=+1999.348728851" watchObservedRunningTime="2025-11-24 07:28:19.45164683 +0000 UTC m=+1999.352238535" Nov 24 07:28:59 crc kubenswrapper[4809]: I1124 07:28:59.358030 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zzdnw"] Nov 24 07:28:59 crc kubenswrapper[4809]: I1124 07:28:59.362065 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zzdnw" Nov 24 07:28:59 crc kubenswrapper[4809]: I1124 07:28:59.371814 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zzdnw"] Nov 24 07:28:59 crc kubenswrapper[4809]: I1124 07:28:59.473028 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88932cc1-662c-4475-9449-744126463ac4-utilities\") pod \"certified-operators-zzdnw\" (UID: \"88932cc1-662c-4475-9449-744126463ac4\") " pod="openshift-marketplace/certified-operators-zzdnw" Nov 24 07:28:59 crc kubenswrapper[4809]: I1124 07:28:59.473115 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5bks\" (UniqueName: \"kubernetes.io/projected/88932cc1-662c-4475-9449-744126463ac4-kube-api-access-s5bks\") pod \"certified-operators-zzdnw\" (UID: \"88932cc1-662c-4475-9449-744126463ac4\") " pod="openshift-marketplace/certified-operators-zzdnw" Nov 24 07:28:59 crc kubenswrapper[4809]: I1124 07:28:59.473323 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88932cc1-662c-4475-9449-744126463ac4-catalog-content\") pod \"certified-operators-zzdnw\" (UID: \"88932cc1-662c-4475-9449-744126463ac4\") " pod="openshift-marketplace/certified-operators-zzdnw" Nov 24 07:28:59 crc kubenswrapper[4809]: I1124 07:28:59.575162 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88932cc1-662c-4475-9449-744126463ac4-catalog-content\") pod \"certified-operators-zzdnw\" (UID: \"88932cc1-662c-4475-9449-744126463ac4\") " pod="openshift-marketplace/certified-operators-zzdnw" Nov 24 07:28:59 crc kubenswrapper[4809]: I1124 07:28:59.575300 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88932cc1-662c-4475-9449-744126463ac4-utilities\") pod \"certified-operators-zzdnw\" (UID: \"88932cc1-662c-4475-9449-744126463ac4\") " pod="openshift-marketplace/certified-operators-zzdnw" Nov 24 07:28:59 crc kubenswrapper[4809]: I1124 07:28:59.575336 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5bks\" (UniqueName: \"kubernetes.io/projected/88932cc1-662c-4475-9449-744126463ac4-kube-api-access-s5bks\") pod \"certified-operators-zzdnw\" (UID: \"88932cc1-662c-4475-9449-744126463ac4\") " pod="openshift-marketplace/certified-operators-zzdnw" Nov 24 07:28:59 crc kubenswrapper[4809]: I1124 07:28:59.575779 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88932cc1-662c-4475-9449-744126463ac4-catalog-content\") pod \"certified-operators-zzdnw\" (UID: \"88932cc1-662c-4475-9449-744126463ac4\") " pod="openshift-marketplace/certified-operators-zzdnw" Nov 24 07:28:59 crc kubenswrapper[4809]: I1124 07:28:59.575806 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88932cc1-662c-4475-9449-744126463ac4-utilities\") pod \"certified-operators-zzdnw\" (UID: \"88932cc1-662c-4475-9449-744126463ac4\") " pod="openshift-marketplace/certified-operators-zzdnw" Nov 24 07:28:59 crc kubenswrapper[4809]: I1124 07:28:59.598583 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5bks\" (UniqueName: \"kubernetes.io/projected/88932cc1-662c-4475-9449-744126463ac4-kube-api-access-s5bks\") pod \"certified-operators-zzdnw\" (UID: \"88932cc1-662c-4475-9449-744126463ac4\") " pod="openshift-marketplace/certified-operators-zzdnw" Nov 24 07:28:59 crc kubenswrapper[4809]: I1124 07:28:59.700672 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zzdnw" Nov 24 07:29:00 crc kubenswrapper[4809]: I1124 07:29:00.197089 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zzdnw"] Nov 24 07:29:00 crc kubenswrapper[4809]: I1124 07:29:00.876358 4809 generic.go:334] "Generic (PLEG): container finished" podID="88932cc1-662c-4475-9449-744126463ac4" containerID="387b12d72ec872911b3d7dda9a9b82765d235c8e867f3c48425d029182a9c5ec" exitCode=0 Nov 24 07:29:00 crc kubenswrapper[4809]: I1124 07:29:00.876414 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzdnw" event={"ID":"88932cc1-662c-4475-9449-744126463ac4","Type":"ContainerDied","Data":"387b12d72ec872911b3d7dda9a9b82765d235c8e867f3c48425d029182a9c5ec"} Nov 24 07:29:00 crc kubenswrapper[4809]: I1124 07:29:00.876443 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzdnw" event={"ID":"88932cc1-662c-4475-9449-744126463ac4","Type":"ContainerStarted","Data":"1ba46e7125cc248f2f380be3602ba3439787a5721be7d58c74143cf32a417f3a"} Nov 24 07:29:02 crc kubenswrapper[4809]: I1124 07:29:02.900720 4809 generic.go:334] "Generic (PLEG): container finished" podID="88932cc1-662c-4475-9449-744126463ac4" containerID="6be29c7c2b21bfadc38625256ece6fb6f4003c648c7afd3f5f6983720b76a644" exitCode=0 Nov 24 07:29:02 crc kubenswrapper[4809]: I1124 07:29:02.912099 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzdnw" event={"ID":"88932cc1-662c-4475-9449-744126463ac4","Type":"ContainerDied","Data":"6be29c7c2b21bfadc38625256ece6fb6f4003c648c7afd3f5f6983720b76a644"} Nov 24 07:29:03 crc kubenswrapper[4809]: I1124 07:29:03.912762 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzdnw" event={"ID":"88932cc1-662c-4475-9449-744126463ac4","Type":"ContainerStarted","Data":"a5c3ff1efc5928682d7e1f5e682f849aaa30276c8d4edadfe65d9b9e8b6d08b6"} Nov 24 07:29:03 crc kubenswrapper[4809]: I1124 07:29:03.944196 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zzdnw" podStartSLOduration=2.4615053270000002 podStartE2EDuration="4.944174677s" podCreationTimestamp="2025-11-24 07:28:59 +0000 UTC" firstStartedPulling="2025-11-24 07:29:00.879189232 +0000 UTC m=+2040.779780967" lastFinishedPulling="2025-11-24 07:29:03.361858612 +0000 UTC m=+2043.262450317" observedRunningTime="2025-11-24 07:29:03.936223805 +0000 UTC m=+2043.836815530" watchObservedRunningTime="2025-11-24 07:29:03.944174677 +0000 UTC m=+2043.844766382" Nov 24 07:29:09 crc kubenswrapper[4809]: I1124 07:29:09.701454 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zzdnw" Nov 24 07:29:09 crc kubenswrapper[4809]: I1124 07:29:09.702286 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zzdnw" Nov 24 07:29:09 crc kubenswrapper[4809]: I1124 07:29:09.760279 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zzdnw" Nov 24 07:29:10 crc kubenswrapper[4809]: I1124 07:29:10.024594 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zzdnw" Nov 24 07:29:10 crc kubenswrapper[4809]: I1124 07:29:10.080674 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zzdnw"] Nov 24 07:29:11 crc kubenswrapper[4809]: I1124 07:29:11.992315 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zzdnw" podUID="88932cc1-662c-4475-9449-744126463ac4" containerName="registry-server" containerID="cri-o://a5c3ff1efc5928682d7e1f5e682f849aaa30276c8d4edadfe65d9b9e8b6d08b6" gracePeriod=2 Nov 24 07:29:12 crc kubenswrapper[4809]: I1124 07:29:12.500509 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zzdnw" Nov 24 07:29:12 crc kubenswrapper[4809]: I1124 07:29:12.689687 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5bks\" (UniqueName: \"kubernetes.io/projected/88932cc1-662c-4475-9449-744126463ac4-kube-api-access-s5bks\") pod \"88932cc1-662c-4475-9449-744126463ac4\" (UID: \"88932cc1-662c-4475-9449-744126463ac4\") " Nov 24 07:29:12 crc kubenswrapper[4809]: I1124 07:29:12.689790 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88932cc1-662c-4475-9449-744126463ac4-utilities\") pod \"88932cc1-662c-4475-9449-744126463ac4\" (UID: \"88932cc1-662c-4475-9449-744126463ac4\") " Nov 24 07:29:12 crc kubenswrapper[4809]: I1124 07:29:12.690413 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88932cc1-662c-4475-9449-744126463ac4-catalog-content\") pod \"88932cc1-662c-4475-9449-744126463ac4\" (UID: \"88932cc1-662c-4475-9449-744126463ac4\") " Nov 24 07:29:12 crc kubenswrapper[4809]: I1124 07:29:12.690912 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88932cc1-662c-4475-9449-744126463ac4-utilities" (OuterVolumeSpecName: "utilities") pod "88932cc1-662c-4475-9449-744126463ac4" (UID: "88932cc1-662c-4475-9449-744126463ac4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:29:12 crc kubenswrapper[4809]: I1124 07:29:12.691373 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88932cc1-662c-4475-9449-744126463ac4-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:29:12 crc kubenswrapper[4809]: I1124 07:29:12.702264 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88932cc1-662c-4475-9449-744126463ac4-kube-api-access-s5bks" (OuterVolumeSpecName: "kube-api-access-s5bks") pod "88932cc1-662c-4475-9449-744126463ac4" (UID: "88932cc1-662c-4475-9449-744126463ac4"). InnerVolumeSpecName "kube-api-access-s5bks". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:29:12 crc kubenswrapper[4809]: I1124 07:29:12.734378 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88932cc1-662c-4475-9449-744126463ac4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "88932cc1-662c-4475-9449-744126463ac4" (UID: "88932cc1-662c-4475-9449-744126463ac4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:29:12 crc kubenswrapper[4809]: I1124 07:29:12.793321 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5bks\" (UniqueName: \"kubernetes.io/projected/88932cc1-662c-4475-9449-744126463ac4-kube-api-access-s5bks\") on node \"crc\" DevicePath \"\"" Nov 24 07:29:12 crc kubenswrapper[4809]: I1124 07:29:12.793365 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88932cc1-662c-4475-9449-744126463ac4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:29:13 crc kubenswrapper[4809]: I1124 07:29:13.005755 4809 generic.go:334] "Generic (PLEG): container finished" podID="88932cc1-662c-4475-9449-744126463ac4" containerID="a5c3ff1efc5928682d7e1f5e682f849aaa30276c8d4edadfe65d9b9e8b6d08b6" exitCode=0 Nov 24 07:29:13 crc kubenswrapper[4809]: I1124 07:29:13.005816 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzdnw" event={"ID":"88932cc1-662c-4475-9449-744126463ac4","Type":"ContainerDied","Data":"a5c3ff1efc5928682d7e1f5e682f849aaa30276c8d4edadfe65d9b9e8b6d08b6"} Nov 24 07:29:13 crc kubenswrapper[4809]: I1124 07:29:13.005825 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zzdnw" Nov 24 07:29:13 crc kubenswrapper[4809]: I1124 07:29:13.005846 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzdnw" event={"ID":"88932cc1-662c-4475-9449-744126463ac4","Type":"ContainerDied","Data":"1ba46e7125cc248f2f380be3602ba3439787a5721be7d58c74143cf32a417f3a"} Nov 24 07:29:13 crc kubenswrapper[4809]: I1124 07:29:13.005864 4809 scope.go:117] "RemoveContainer" containerID="a5c3ff1efc5928682d7e1f5e682f849aaa30276c8d4edadfe65d9b9e8b6d08b6" Nov 24 07:29:13 crc kubenswrapper[4809]: I1124 07:29:13.036162 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zzdnw"] Nov 24 07:29:13 crc kubenswrapper[4809]: I1124 07:29:13.044986 4809 scope.go:117] "RemoveContainer" containerID="6be29c7c2b21bfadc38625256ece6fb6f4003c648c7afd3f5f6983720b76a644" Nov 24 07:29:13 crc kubenswrapper[4809]: I1124 07:29:13.045408 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zzdnw"] Nov 24 07:29:13 crc kubenswrapper[4809]: I1124 07:29:13.078183 4809 scope.go:117] "RemoveContainer" containerID="387b12d72ec872911b3d7dda9a9b82765d235c8e867f3c48425d029182a9c5ec" Nov 24 07:29:13 crc kubenswrapper[4809]: I1124 07:29:13.116666 4809 scope.go:117] "RemoveContainer" containerID="a5c3ff1efc5928682d7e1f5e682f849aaa30276c8d4edadfe65d9b9e8b6d08b6" Nov 24 07:29:13 crc kubenswrapper[4809]: E1124 07:29:13.117583 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5c3ff1efc5928682d7e1f5e682f849aaa30276c8d4edadfe65d9b9e8b6d08b6\": container with ID starting with a5c3ff1efc5928682d7e1f5e682f849aaa30276c8d4edadfe65d9b9e8b6d08b6 not found: ID does not exist" containerID="a5c3ff1efc5928682d7e1f5e682f849aaa30276c8d4edadfe65d9b9e8b6d08b6" Nov 24 07:29:13 crc kubenswrapper[4809]: I1124 07:29:13.117625 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5c3ff1efc5928682d7e1f5e682f849aaa30276c8d4edadfe65d9b9e8b6d08b6"} err="failed to get container status \"a5c3ff1efc5928682d7e1f5e682f849aaa30276c8d4edadfe65d9b9e8b6d08b6\": rpc error: code = NotFound desc = could not find container \"a5c3ff1efc5928682d7e1f5e682f849aaa30276c8d4edadfe65d9b9e8b6d08b6\": container with ID starting with a5c3ff1efc5928682d7e1f5e682f849aaa30276c8d4edadfe65d9b9e8b6d08b6 not found: ID does not exist" Nov 24 07:29:13 crc kubenswrapper[4809]: I1124 07:29:13.117652 4809 scope.go:117] "RemoveContainer" containerID="6be29c7c2b21bfadc38625256ece6fb6f4003c648c7afd3f5f6983720b76a644" Nov 24 07:29:13 crc kubenswrapper[4809]: E1124 07:29:13.117935 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6be29c7c2b21bfadc38625256ece6fb6f4003c648c7afd3f5f6983720b76a644\": container with ID starting with 6be29c7c2b21bfadc38625256ece6fb6f4003c648c7afd3f5f6983720b76a644 not found: ID does not exist" containerID="6be29c7c2b21bfadc38625256ece6fb6f4003c648c7afd3f5f6983720b76a644" Nov 24 07:29:13 crc kubenswrapper[4809]: I1124 07:29:13.117981 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6be29c7c2b21bfadc38625256ece6fb6f4003c648c7afd3f5f6983720b76a644"} err="failed to get container status \"6be29c7c2b21bfadc38625256ece6fb6f4003c648c7afd3f5f6983720b76a644\": rpc error: code = NotFound desc = could not find container \"6be29c7c2b21bfadc38625256ece6fb6f4003c648c7afd3f5f6983720b76a644\": container with ID starting with 6be29c7c2b21bfadc38625256ece6fb6f4003c648c7afd3f5f6983720b76a644 not found: ID does not exist" Nov 24 07:29:13 crc kubenswrapper[4809]: I1124 07:29:13.118000 4809 scope.go:117] "RemoveContainer" containerID="387b12d72ec872911b3d7dda9a9b82765d235c8e867f3c48425d029182a9c5ec" Nov 24 07:29:13 crc kubenswrapper[4809]: E1124 07:29:13.118248 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"387b12d72ec872911b3d7dda9a9b82765d235c8e867f3c48425d029182a9c5ec\": container with ID starting with 387b12d72ec872911b3d7dda9a9b82765d235c8e867f3c48425d029182a9c5ec not found: ID does not exist" containerID="387b12d72ec872911b3d7dda9a9b82765d235c8e867f3c48425d029182a9c5ec" Nov 24 07:29:13 crc kubenswrapper[4809]: I1124 07:29:13.118277 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"387b12d72ec872911b3d7dda9a9b82765d235c8e867f3c48425d029182a9c5ec"} err="failed to get container status \"387b12d72ec872911b3d7dda9a9b82765d235c8e867f3c48425d029182a9c5ec\": rpc error: code = NotFound desc = could not find container \"387b12d72ec872911b3d7dda9a9b82765d235c8e867f3c48425d029182a9c5ec\": container with ID starting with 387b12d72ec872911b3d7dda9a9b82765d235c8e867f3c48425d029182a9c5ec not found: ID does not exist" Nov 24 07:29:14 crc kubenswrapper[4809]: I1124 07:29:14.911530 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88932cc1-662c-4475-9449-744126463ac4" path="/var/lib/kubelet/pods/88932cc1-662c-4475-9449-744126463ac4/volumes" Nov 24 07:29:52 crc kubenswrapper[4809]: I1124 07:29:52.336177 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zgzgz"] Nov 24 07:29:52 crc kubenswrapper[4809]: E1124 07:29:52.337050 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88932cc1-662c-4475-9449-744126463ac4" containerName="extract-utilities" Nov 24 07:29:52 crc kubenswrapper[4809]: I1124 07:29:52.337063 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="88932cc1-662c-4475-9449-744126463ac4" containerName="extract-utilities" Nov 24 07:29:52 crc kubenswrapper[4809]: E1124 07:29:52.337092 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88932cc1-662c-4475-9449-744126463ac4" containerName="extract-content" Nov 24 07:29:52 crc kubenswrapper[4809]: I1124 07:29:52.337098 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="88932cc1-662c-4475-9449-744126463ac4" containerName="extract-content" Nov 24 07:29:52 crc kubenswrapper[4809]: E1124 07:29:52.337120 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88932cc1-662c-4475-9449-744126463ac4" containerName="registry-server" Nov 24 07:29:52 crc kubenswrapper[4809]: I1124 07:29:52.337125 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="88932cc1-662c-4475-9449-744126463ac4" containerName="registry-server" Nov 24 07:29:52 crc kubenswrapper[4809]: I1124 07:29:52.337298 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="88932cc1-662c-4475-9449-744126463ac4" containerName="registry-server" Nov 24 07:29:52 crc kubenswrapper[4809]: I1124 07:29:52.338538 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zgzgz" Nov 24 07:29:52 crc kubenswrapper[4809]: I1124 07:29:52.365130 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zgzgz"] Nov 24 07:29:52 crc kubenswrapper[4809]: I1124 07:29:52.470473 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd7588f7-f400-41e6-a14e-d01e548130a3-catalog-content\") pod \"redhat-marketplace-zgzgz\" (UID: \"cd7588f7-f400-41e6-a14e-d01e548130a3\") " pod="openshift-marketplace/redhat-marketplace-zgzgz" Nov 24 07:29:52 crc kubenswrapper[4809]: I1124 07:29:52.470603 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd7588f7-f400-41e6-a14e-d01e548130a3-utilities\") pod \"redhat-marketplace-zgzgz\" (UID: \"cd7588f7-f400-41e6-a14e-d01e548130a3\") " pod="openshift-marketplace/redhat-marketplace-zgzgz" Nov 24 07:29:52 crc kubenswrapper[4809]: I1124 07:29:52.470688 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dr78w\" (UniqueName: \"kubernetes.io/projected/cd7588f7-f400-41e6-a14e-d01e548130a3-kube-api-access-dr78w\") pod \"redhat-marketplace-zgzgz\" (UID: \"cd7588f7-f400-41e6-a14e-d01e548130a3\") " pod="openshift-marketplace/redhat-marketplace-zgzgz" Nov 24 07:29:52 crc kubenswrapper[4809]: I1124 07:29:52.572561 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dr78w\" (UniqueName: \"kubernetes.io/projected/cd7588f7-f400-41e6-a14e-d01e548130a3-kube-api-access-dr78w\") pod \"redhat-marketplace-zgzgz\" (UID: \"cd7588f7-f400-41e6-a14e-d01e548130a3\") " pod="openshift-marketplace/redhat-marketplace-zgzgz" Nov 24 07:29:52 crc kubenswrapper[4809]: I1124 07:29:52.572957 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd7588f7-f400-41e6-a14e-d01e548130a3-catalog-content\") pod \"redhat-marketplace-zgzgz\" (UID: \"cd7588f7-f400-41e6-a14e-d01e548130a3\") " pod="openshift-marketplace/redhat-marketplace-zgzgz" Nov 24 07:29:52 crc kubenswrapper[4809]: I1124 07:29:52.573172 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd7588f7-f400-41e6-a14e-d01e548130a3-utilities\") pod \"redhat-marketplace-zgzgz\" (UID: \"cd7588f7-f400-41e6-a14e-d01e548130a3\") " pod="openshift-marketplace/redhat-marketplace-zgzgz" Nov 24 07:29:52 crc kubenswrapper[4809]: I1124 07:29:52.573519 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd7588f7-f400-41e6-a14e-d01e548130a3-catalog-content\") pod \"redhat-marketplace-zgzgz\" (UID: \"cd7588f7-f400-41e6-a14e-d01e548130a3\") " pod="openshift-marketplace/redhat-marketplace-zgzgz" Nov 24 07:29:52 crc kubenswrapper[4809]: I1124 07:29:52.573678 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd7588f7-f400-41e6-a14e-d01e548130a3-utilities\") pod \"redhat-marketplace-zgzgz\" (UID: \"cd7588f7-f400-41e6-a14e-d01e548130a3\") " pod="openshift-marketplace/redhat-marketplace-zgzgz" Nov 24 07:29:52 crc kubenswrapper[4809]: I1124 07:29:52.592687 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dr78w\" (UniqueName: \"kubernetes.io/projected/cd7588f7-f400-41e6-a14e-d01e548130a3-kube-api-access-dr78w\") pod \"redhat-marketplace-zgzgz\" (UID: \"cd7588f7-f400-41e6-a14e-d01e548130a3\") " pod="openshift-marketplace/redhat-marketplace-zgzgz" Nov 24 07:29:52 crc kubenswrapper[4809]: I1124 07:29:52.721144 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zgzgz" Nov 24 07:29:53 crc kubenswrapper[4809]: I1124 07:29:53.184685 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zgzgz"] Nov 24 07:29:53 crc kubenswrapper[4809]: I1124 07:29:53.437435 4809 generic.go:334] "Generic (PLEG): container finished" podID="cd7588f7-f400-41e6-a14e-d01e548130a3" containerID="bb871ab8c8689101f5cfafeb11bce1d381c028e34e7f000676150415a934dc90" exitCode=0 Nov 24 07:29:53 crc kubenswrapper[4809]: I1124 07:29:53.437738 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zgzgz" event={"ID":"cd7588f7-f400-41e6-a14e-d01e548130a3","Type":"ContainerDied","Data":"bb871ab8c8689101f5cfafeb11bce1d381c028e34e7f000676150415a934dc90"} Nov 24 07:29:53 crc kubenswrapper[4809]: I1124 07:29:53.439248 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zgzgz" event={"ID":"cd7588f7-f400-41e6-a14e-d01e548130a3","Type":"ContainerStarted","Data":"a9cc79aa32df714a0505485f4667cb737b2e542c28c18306acbb1699d490ffb0"} Nov 24 07:29:54 crc kubenswrapper[4809]: I1124 07:29:54.450079 4809 generic.go:334] "Generic (PLEG): container finished" podID="cd7588f7-f400-41e6-a14e-d01e548130a3" containerID="0aa656f08e00a43f7dedda858966553d04e76f2a6a1d1c93dd4c561146f42ee4" exitCode=0 Nov 24 07:29:54 crc kubenswrapper[4809]: I1124 07:29:54.450217 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zgzgz" event={"ID":"cd7588f7-f400-41e6-a14e-d01e548130a3","Type":"ContainerDied","Data":"0aa656f08e00a43f7dedda858966553d04e76f2a6a1d1c93dd4c561146f42ee4"} Nov 24 07:29:55 crc kubenswrapper[4809]: I1124 07:29:55.461105 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zgzgz" event={"ID":"cd7588f7-f400-41e6-a14e-d01e548130a3","Type":"ContainerStarted","Data":"16a25a721de8f2e7e8aabce6d8daeda0a656049b6375b1ebdc94c92dc183ad00"} Nov 24 07:29:55 crc kubenswrapper[4809]: I1124 07:29:55.479991 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zgzgz" podStartSLOduration=2.093459822 podStartE2EDuration="3.479956377s" podCreationTimestamp="2025-11-24 07:29:52 +0000 UTC" firstStartedPulling="2025-11-24 07:29:53.439951869 +0000 UTC m=+2093.340543574" lastFinishedPulling="2025-11-24 07:29:54.826448394 +0000 UTC m=+2094.727040129" observedRunningTime="2025-11-24 07:29:55.476501235 +0000 UTC m=+2095.377092950" watchObservedRunningTime="2025-11-24 07:29:55.479956377 +0000 UTC m=+2095.380548082" Nov 24 07:30:00 crc kubenswrapper[4809]: I1124 07:30:00.151016 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399490-fkmkm"] Nov 24 07:30:00 crc kubenswrapper[4809]: I1124 07:30:00.153773 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-fkmkm" Nov 24 07:30:00 crc kubenswrapper[4809]: I1124 07:30:00.155715 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 07:30:00 crc kubenswrapper[4809]: I1124 07:30:00.156409 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 07:30:00 crc kubenswrapper[4809]: I1124 07:30:00.165190 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399490-fkmkm"] Nov 24 07:30:00 crc kubenswrapper[4809]: I1124 07:30:00.221856 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5f0f244-c1f0-4c1c-84ef-9c994b349739-config-volume\") pod \"collect-profiles-29399490-fkmkm\" (UID: \"a5f0f244-c1f0-4c1c-84ef-9c994b349739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-fkmkm" Nov 24 07:30:00 crc kubenswrapper[4809]: I1124 07:30:00.221915 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5f0f244-c1f0-4c1c-84ef-9c994b349739-secret-volume\") pod \"collect-profiles-29399490-fkmkm\" (UID: \"a5f0f244-c1f0-4c1c-84ef-9c994b349739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-fkmkm" Nov 24 07:30:00 crc kubenswrapper[4809]: I1124 07:30:00.222094 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rhnt\" (UniqueName: \"kubernetes.io/projected/a5f0f244-c1f0-4c1c-84ef-9c994b349739-kube-api-access-8rhnt\") pod \"collect-profiles-29399490-fkmkm\" (UID: \"a5f0f244-c1f0-4c1c-84ef-9c994b349739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-fkmkm" Nov 24 07:30:00 crc kubenswrapper[4809]: I1124 07:30:00.323248 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5f0f244-c1f0-4c1c-84ef-9c994b349739-config-volume\") pod \"collect-profiles-29399490-fkmkm\" (UID: \"a5f0f244-c1f0-4c1c-84ef-9c994b349739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-fkmkm" Nov 24 07:30:00 crc kubenswrapper[4809]: I1124 07:30:00.323299 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5f0f244-c1f0-4c1c-84ef-9c994b349739-secret-volume\") pod \"collect-profiles-29399490-fkmkm\" (UID: \"a5f0f244-c1f0-4c1c-84ef-9c994b349739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-fkmkm" Nov 24 07:30:00 crc kubenswrapper[4809]: I1124 07:30:00.323399 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rhnt\" (UniqueName: \"kubernetes.io/projected/a5f0f244-c1f0-4c1c-84ef-9c994b349739-kube-api-access-8rhnt\") pod \"collect-profiles-29399490-fkmkm\" (UID: \"a5f0f244-c1f0-4c1c-84ef-9c994b349739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-fkmkm" Nov 24 07:30:00 crc kubenswrapper[4809]: I1124 07:30:00.324948 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5f0f244-c1f0-4c1c-84ef-9c994b349739-config-volume\") pod \"collect-profiles-29399490-fkmkm\" (UID: \"a5f0f244-c1f0-4c1c-84ef-9c994b349739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-fkmkm" Nov 24 07:30:00 crc kubenswrapper[4809]: I1124 07:30:00.336138 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5f0f244-c1f0-4c1c-84ef-9c994b349739-secret-volume\") pod \"collect-profiles-29399490-fkmkm\" (UID: \"a5f0f244-c1f0-4c1c-84ef-9c994b349739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-fkmkm" Nov 24 07:30:00 crc kubenswrapper[4809]: I1124 07:30:00.340173 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rhnt\" (UniqueName: \"kubernetes.io/projected/a5f0f244-c1f0-4c1c-84ef-9c994b349739-kube-api-access-8rhnt\") pod \"collect-profiles-29399490-fkmkm\" (UID: \"a5f0f244-c1f0-4c1c-84ef-9c994b349739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-fkmkm" Nov 24 07:30:00 crc kubenswrapper[4809]: I1124 07:30:00.480569 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-fkmkm" Nov 24 07:30:00 crc kubenswrapper[4809]: I1124 07:30:00.926461 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399490-fkmkm"] Nov 24 07:30:01 crc kubenswrapper[4809]: I1124 07:30:01.571865 4809 generic.go:334] "Generic (PLEG): container finished" podID="a5f0f244-c1f0-4c1c-84ef-9c994b349739" containerID="13c1f8c5cbceba1f6a06d0a004f805e33af3a24fd9b961706842eb7cf99bbb71" exitCode=0 Nov 24 07:30:01 crc kubenswrapper[4809]: I1124 07:30:01.572083 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-fkmkm" event={"ID":"a5f0f244-c1f0-4c1c-84ef-9c994b349739","Type":"ContainerDied","Data":"13c1f8c5cbceba1f6a06d0a004f805e33af3a24fd9b961706842eb7cf99bbb71"} Nov 24 07:30:01 crc kubenswrapper[4809]: I1124 07:30:01.572577 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-fkmkm" event={"ID":"a5f0f244-c1f0-4c1c-84ef-9c994b349739","Type":"ContainerStarted","Data":"a5f1979258025015a42edbcfc2e159c2697f3f51f91003b51debd236372c878d"} Nov 24 07:30:02 crc kubenswrapper[4809]: I1124 07:30:02.721677 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zgzgz" Nov 24 07:30:02 crc kubenswrapper[4809]: I1124 07:30:02.722059 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zgzgz" Nov 24 07:30:02 crc kubenswrapper[4809]: I1124 07:30:02.785081 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zgzgz" Nov 24 07:30:02 crc kubenswrapper[4809]: I1124 07:30:02.891018 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-fkmkm" Nov 24 07:30:02 crc kubenswrapper[4809]: I1124 07:30:02.998690 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rhnt\" (UniqueName: \"kubernetes.io/projected/a5f0f244-c1f0-4c1c-84ef-9c994b349739-kube-api-access-8rhnt\") pod \"a5f0f244-c1f0-4c1c-84ef-9c994b349739\" (UID: \"a5f0f244-c1f0-4c1c-84ef-9c994b349739\") " Nov 24 07:30:02 crc kubenswrapper[4809]: I1124 07:30:02.999116 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5f0f244-c1f0-4c1c-84ef-9c994b349739-secret-volume\") pod \"a5f0f244-c1f0-4c1c-84ef-9c994b349739\" (UID: \"a5f0f244-c1f0-4c1c-84ef-9c994b349739\") " Nov 24 07:30:02 crc kubenswrapper[4809]: I1124 07:30:02.999294 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5f0f244-c1f0-4c1c-84ef-9c994b349739-config-volume\") pod \"a5f0f244-c1f0-4c1c-84ef-9c994b349739\" (UID: \"a5f0f244-c1f0-4c1c-84ef-9c994b349739\") " Nov 24 07:30:03 crc kubenswrapper[4809]: I1124 07:30:03.000262 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5f0f244-c1f0-4c1c-84ef-9c994b349739-config-volume" (OuterVolumeSpecName: "config-volume") pod "a5f0f244-c1f0-4c1c-84ef-9c994b349739" (UID: "a5f0f244-c1f0-4c1c-84ef-9c994b349739"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:30:03 crc kubenswrapper[4809]: I1124 07:30:03.003922 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5f0f244-c1f0-4c1c-84ef-9c994b349739-kube-api-access-8rhnt" (OuterVolumeSpecName: "kube-api-access-8rhnt") pod "a5f0f244-c1f0-4c1c-84ef-9c994b349739" (UID: "a5f0f244-c1f0-4c1c-84ef-9c994b349739"). InnerVolumeSpecName "kube-api-access-8rhnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:30:03 crc kubenswrapper[4809]: I1124 07:30:03.005345 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5f0f244-c1f0-4c1c-84ef-9c994b349739-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a5f0f244-c1f0-4c1c-84ef-9c994b349739" (UID: "a5f0f244-c1f0-4c1c-84ef-9c994b349739"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:30:03 crc kubenswrapper[4809]: I1124 07:30:03.103179 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rhnt\" (UniqueName: \"kubernetes.io/projected/a5f0f244-c1f0-4c1c-84ef-9c994b349739-kube-api-access-8rhnt\") on node \"crc\" DevicePath \"\"" Nov 24 07:30:03 crc kubenswrapper[4809]: I1124 07:30:03.103233 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5f0f244-c1f0-4c1c-84ef-9c994b349739-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 07:30:03 crc kubenswrapper[4809]: I1124 07:30:03.103251 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5f0f244-c1f0-4c1c-84ef-9c994b349739-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 07:30:03 crc kubenswrapper[4809]: I1124 07:30:03.592137 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-fkmkm" Nov 24 07:30:03 crc kubenswrapper[4809]: I1124 07:30:03.592144 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-fkmkm" event={"ID":"a5f0f244-c1f0-4c1c-84ef-9c994b349739","Type":"ContainerDied","Data":"a5f1979258025015a42edbcfc2e159c2697f3f51f91003b51debd236372c878d"} Nov 24 07:30:03 crc kubenswrapper[4809]: I1124 07:30:03.592207 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5f1979258025015a42edbcfc2e159c2697f3f51f91003b51debd236372c878d" Nov 24 07:30:03 crc kubenswrapper[4809]: I1124 07:30:03.650616 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zgzgz" Nov 24 07:30:03 crc kubenswrapper[4809]: I1124 07:30:03.713939 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zgzgz"] Nov 24 07:30:03 crc kubenswrapper[4809]: I1124 07:30:03.980581 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399445-bkt2p"] Nov 24 07:30:03 crc kubenswrapper[4809]: I1124 07:30:03.986763 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399445-bkt2p"] Nov 24 07:30:04 crc kubenswrapper[4809]: I1124 07:30:04.905685 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="476eceb4-f8fa-49c1-9015-5e067998af80" path="/var/lib/kubelet/pods/476eceb4-f8fa-49c1-9015-5e067998af80/volumes" Nov 24 07:30:05 crc kubenswrapper[4809]: I1124 07:30:05.613929 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zgzgz" podUID="cd7588f7-f400-41e6-a14e-d01e548130a3" containerName="registry-server" containerID="cri-o://16a25a721de8f2e7e8aabce6d8daeda0a656049b6375b1ebdc94c92dc183ad00" gracePeriod=2 Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.106112 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zgzgz" Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.266344 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd7588f7-f400-41e6-a14e-d01e548130a3-catalog-content\") pod \"cd7588f7-f400-41e6-a14e-d01e548130a3\" (UID: \"cd7588f7-f400-41e6-a14e-d01e548130a3\") " Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.267046 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dr78w\" (UniqueName: \"kubernetes.io/projected/cd7588f7-f400-41e6-a14e-d01e548130a3-kube-api-access-dr78w\") pod \"cd7588f7-f400-41e6-a14e-d01e548130a3\" (UID: \"cd7588f7-f400-41e6-a14e-d01e548130a3\") " Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.267273 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd7588f7-f400-41e6-a14e-d01e548130a3-utilities\") pod \"cd7588f7-f400-41e6-a14e-d01e548130a3\" (UID: \"cd7588f7-f400-41e6-a14e-d01e548130a3\") " Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.268272 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd7588f7-f400-41e6-a14e-d01e548130a3-utilities" (OuterVolumeSpecName: "utilities") pod "cd7588f7-f400-41e6-a14e-d01e548130a3" (UID: "cd7588f7-f400-41e6-a14e-d01e548130a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.272852 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd7588f7-f400-41e6-a14e-d01e548130a3-kube-api-access-dr78w" (OuterVolumeSpecName: "kube-api-access-dr78w") pod "cd7588f7-f400-41e6-a14e-d01e548130a3" (UID: "cd7588f7-f400-41e6-a14e-d01e548130a3"). InnerVolumeSpecName "kube-api-access-dr78w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.289339 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd7588f7-f400-41e6-a14e-d01e548130a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cd7588f7-f400-41e6-a14e-d01e548130a3" (UID: "cd7588f7-f400-41e6-a14e-d01e548130a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.369837 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dr78w\" (UniqueName: \"kubernetes.io/projected/cd7588f7-f400-41e6-a14e-d01e548130a3-kube-api-access-dr78w\") on node \"crc\" DevicePath \"\"" Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.369889 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd7588f7-f400-41e6-a14e-d01e548130a3-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.369903 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd7588f7-f400-41e6-a14e-d01e548130a3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.626377 4809 generic.go:334] "Generic (PLEG): container finished" podID="cd7588f7-f400-41e6-a14e-d01e548130a3" containerID="16a25a721de8f2e7e8aabce6d8daeda0a656049b6375b1ebdc94c92dc183ad00" exitCode=0 Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.626427 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zgzgz" event={"ID":"cd7588f7-f400-41e6-a14e-d01e548130a3","Type":"ContainerDied","Data":"16a25a721de8f2e7e8aabce6d8daeda0a656049b6375b1ebdc94c92dc183ad00"} Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.626463 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zgzgz" event={"ID":"cd7588f7-f400-41e6-a14e-d01e548130a3","Type":"ContainerDied","Data":"a9cc79aa32df714a0505485f4667cb737b2e542c28c18306acbb1699d490ffb0"} Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.626486 4809 scope.go:117] "RemoveContainer" containerID="16a25a721de8f2e7e8aabce6d8daeda0a656049b6375b1ebdc94c92dc183ad00" Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.626560 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zgzgz" Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.661312 4809 scope.go:117] "RemoveContainer" containerID="0aa656f08e00a43f7dedda858966553d04e76f2a6a1d1c93dd4c561146f42ee4" Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.686866 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zgzgz"] Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.703983 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zgzgz"] Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.710194 4809 scope.go:117] "RemoveContainer" containerID="bb871ab8c8689101f5cfafeb11bce1d381c028e34e7f000676150415a934dc90" Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.771512 4809 scope.go:117] "RemoveContainer" containerID="16a25a721de8f2e7e8aabce6d8daeda0a656049b6375b1ebdc94c92dc183ad00" Nov 24 07:30:06 crc kubenswrapper[4809]: E1124 07:30:06.772435 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16a25a721de8f2e7e8aabce6d8daeda0a656049b6375b1ebdc94c92dc183ad00\": container with ID starting with 16a25a721de8f2e7e8aabce6d8daeda0a656049b6375b1ebdc94c92dc183ad00 not found: ID does not exist" containerID="16a25a721de8f2e7e8aabce6d8daeda0a656049b6375b1ebdc94c92dc183ad00" Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.772489 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16a25a721de8f2e7e8aabce6d8daeda0a656049b6375b1ebdc94c92dc183ad00"} err="failed to get container status \"16a25a721de8f2e7e8aabce6d8daeda0a656049b6375b1ebdc94c92dc183ad00\": rpc error: code = NotFound desc = could not find container \"16a25a721de8f2e7e8aabce6d8daeda0a656049b6375b1ebdc94c92dc183ad00\": container with ID starting with 16a25a721de8f2e7e8aabce6d8daeda0a656049b6375b1ebdc94c92dc183ad00 not found: ID does not exist" Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.772522 4809 scope.go:117] "RemoveContainer" containerID="0aa656f08e00a43f7dedda858966553d04e76f2a6a1d1c93dd4c561146f42ee4" Nov 24 07:30:06 crc kubenswrapper[4809]: E1124 07:30:06.772874 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0aa656f08e00a43f7dedda858966553d04e76f2a6a1d1c93dd4c561146f42ee4\": container with ID starting with 0aa656f08e00a43f7dedda858966553d04e76f2a6a1d1c93dd4c561146f42ee4 not found: ID does not exist" containerID="0aa656f08e00a43f7dedda858966553d04e76f2a6a1d1c93dd4c561146f42ee4" Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.772921 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0aa656f08e00a43f7dedda858966553d04e76f2a6a1d1c93dd4c561146f42ee4"} err="failed to get container status \"0aa656f08e00a43f7dedda858966553d04e76f2a6a1d1c93dd4c561146f42ee4\": rpc error: code = NotFound desc = could not find container \"0aa656f08e00a43f7dedda858966553d04e76f2a6a1d1c93dd4c561146f42ee4\": container with ID starting with 0aa656f08e00a43f7dedda858966553d04e76f2a6a1d1c93dd4c561146f42ee4 not found: ID does not exist" Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.772950 4809 scope.go:117] "RemoveContainer" containerID="bb871ab8c8689101f5cfafeb11bce1d381c028e34e7f000676150415a934dc90" Nov 24 07:30:06 crc kubenswrapper[4809]: E1124 07:30:06.773237 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb871ab8c8689101f5cfafeb11bce1d381c028e34e7f000676150415a934dc90\": container with ID starting with bb871ab8c8689101f5cfafeb11bce1d381c028e34e7f000676150415a934dc90 not found: ID does not exist" containerID="bb871ab8c8689101f5cfafeb11bce1d381c028e34e7f000676150415a934dc90" Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.773268 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb871ab8c8689101f5cfafeb11bce1d381c028e34e7f000676150415a934dc90"} err="failed to get container status \"bb871ab8c8689101f5cfafeb11bce1d381c028e34e7f000676150415a934dc90\": rpc error: code = NotFound desc = could not find container \"bb871ab8c8689101f5cfafeb11bce1d381c028e34e7f000676150415a934dc90\": container with ID starting with bb871ab8c8689101f5cfafeb11bce1d381c028e34e7f000676150415a934dc90 not found: ID does not exist" Nov 24 07:30:06 crc kubenswrapper[4809]: I1124 07:30:06.903230 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd7588f7-f400-41e6-a14e-d01e548130a3" path="/var/lib/kubelet/pods/cd7588f7-f400-41e6-a14e-d01e548130a3/volumes" Nov 24 07:30:10 crc kubenswrapper[4809]: I1124 07:30:10.032907 4809 scope.go:117] "RemoveContainer" containerID="7c9151d6d63096bfd5e884d4181017e2d336422e70c5bf3faa413167e62b0b76" Nov 24 07:30:18 crc kubenswrapper[4809]: I1124 07:30:18.043648 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:30:18 crc kubenswrapper[4809]: I1124 07:30:18.044278 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:30:48 crc kubenswrapper[4809]: I1124 07:30:48.043919 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:30:48 crc kubenswrapper[4809]: I1124 07:30:48.044877 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:31:18 crc kubenswrapper[4809]: I1124 07:31:18.044138 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:31:18 crc kubenswrapper[4809]: I1124 07:31:18.044743 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:31:18 crc kubenswrapper[4809]: I1124 07:31:18.044806 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 07:31:18 crc kubenswrapper[4809]: I1124 07:31:18.045672 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add"} pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:31:18 crc kubenswrapper[4809]: I1124 07:31:18.045765 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" containerID="cri-o://b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" gracePeriod=600 Nov 24 07:31:18 crc kubenswrapper[4809]: E1124 07:31:18.167764 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:31:18 crc kubenswrapper[4809]: I1124 07:31:18.338907 4809 generic.go:334] "Generic (PLEG): container finished" podID="be8050d5-1aea-453e-a06e-87702b0e856b" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" exitCode=0 Nov 24 07:31:18 crc kubenswrapper[4809]: I1124 07:31:18.338948 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerDied","Data":"b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add"} Nov 24 07:31:18 crc kubenswrapper[4809]: I1124 07:31:18.338984 4809 scope.go:117] "RemoveContainer" containerID="da763efb03a569decc0792fe26fc2e80b6b2ae304146cd9bddaf55c16474daac" Nov 24 07:31:18 crc kubenswrapper[4809]: I1124 07:31:18.339773 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:31:18 crc kubenswrapper[4809]: E1124 07:31:18.340105 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:31:31 crc kubenswrapper[4809]: I1124 07:31:31.891230 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:31:31 crc kubenswrapper[4809]: E1124 07:31:31.891995 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:31:43 crc kubenswrapper[4809]: I1124 07:31:43.890895 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:31:43 crc kubenswrapper[4809]: E1124 07:31:43.891701 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:31:54 crc kubenswrapper[4809]: I1124 07:31:54.891960 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:31:54 crc kubenswrapper[4809]: E1124 07:31:54.893138 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:32:07 crc kubenswrapper[4809]: I1124 07:32:07.892550 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:32:07 crc kubenswrapper[4809]: E1124 07:32:07.893781 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:32:21 crc kubenswrapper[4809]: I1124 07:32:21.891778 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:32:21 crc kubenswrapper[4809]: E1124 07:32:21.893105 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:32:36 crc kubenswrapper[4809]: I1124 07:32:36.891924 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:32:36 crc kubenswrapper[4809]: E1124 07:32:36.892963 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:32:48 crc kubenswrapper[4809]: I1124 07:32:48.891889 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:32:48 crc kubenswrapper[4809]: E1124 07:32:48.892647 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:32:57 crc kubenswrapper[4809]: I1124 07:32:57.351534 4809 generic.go:334] "Generic (PLEG): container finished" podID="fd55a20b-bdd2-40bf-85d7-c0b485352b8f" containerID="83dd7678b8b5ef6575cc667a269d20ce80314a351a65815e6eb46f83d56c9e3f" exitCode=0 Nov 24 07:32:57 crc kubenswrapper[4809]: I1124 07:32:57.351598 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" event={"ID":"fd55a20b-bdd2-40bf-85d7-c0b485352b8f","Type":"ContainerDied","Data":"83dd7678b8b5ef6575cc667a269d20ce80314a351a65815e6eb46f83d56c9e3f"} Nov 24 07:32:58 crc kubenswrapper[4809]: I1124 07:32:58.773530 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" Nov 24 07:32:58 crc kubenswrapper[4809]: I1124 07:32:58.896928 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-libvirt-combined-ca-bundle\") pod \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\" (UID: \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\") " Nov 24 07:32:58 crc kubenswrapper[4809]: I1124 07:32:58.897120 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-inventory\") pod \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\" (UID: \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\") " Nov 24 07:32:58 crc kubenswrapper[4809]: I1124 07:32:58.897147 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-libvirt-secret-0\") pod \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\" (UID: \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\") " Nov 24 07:32:58 crc kubenswrapper[4809]: I1124 07:32:58.897183 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-ssh-key\") pod \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\" (UID: \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\") " Nov 24 07:32:58 crc kubenswrapper[4809]: I1124 07:32:58.897253 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggxb4\" (UniqueName: \"kubernetes.io/projected/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-kube-api-access-ggxb4\") pod \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\" (UID: \"fd55a20b-bdd2-40bf-85d7-c0b485352b8f\") " Nov 24 07:32:58 crc kubenswrapper[4809]: I1124 07:32:58.904038 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "fd55a20b-bdd2-40bf-85d7-c0b485352b8f" (UID: "fd55a20b-bdd2-40bf-85d7-c0b485352b8f"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:32:58 crc kubenswrapper[4809]: I1124 07:32:58.904218 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-kube-api-access-ggxb4" (OuterVolumeSpecName: "kube-api-access-ggxb4") pod "fd55a20b-bdd2-40bf-85d7-c0b485352b8f" (UID: "fd55a20b-bdd2-40bf-85d7-c0b485352b8f"). InnerVolumeSpecName "kube-api-access-ggxb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:32:58 crc kubenswrapper[4809]: I1124 07:32:58.923888 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "fd55a20b-bdd2-40bf-85d7-c0b485352b8f" (UID: "fd55a20b-bdd2-40bf-85d7-c0b485352b8f"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:32:58 crc kubenswrapper[4809]: I1124 07:32:58.928114 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fd55a20b-bdd2-40bf-85d7-c0b485352b8f" (UID: "fd55a20b-bdd2-40bf-85d7-c0b485352b8f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:32:58 crc kubenswrapper[4809]: I1124 07:32:58.938343 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-inventory" (OuterVolumeSpecName: "inventory") pod "fd55a20b-bdd2-40bf-85d7-c0b485352b8f" (UID: "fd55a20b-bdd2-40bf-85d7-c0b485352b8f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:32:58 crc kubenswrapper[4809]: I1124 07:32:58.999434 4809 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:32:58 crc kubenswrapper[4809]: I1124 07:32:58.999468 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 07:32:58 crc kubenswrapper[4809]: I1124 07:32:58.999480 4809 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:32:58 crc kubenswrapper[4809]: I1124 07:32:58.999493 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 07:32:58 crc kubenswrapper[4809]: I1124 07:32:58.999506 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggxb4\" (UniqueName: \"kubernetes.io/projected/fd55a20b-bdd2-40bf-85d7-c0b485352b8f-kube-api-access-ggxb4\") on node \"crc\" DevicePath \"\"" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.375866 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" event={"ID":"fd55a20b-bdd2-40bf-85d7-c0b485352b8f","Type":"ContainerDied","Data":"5836f266843a9c18248bacd679bf75921ccbdd00f3b48a8df366ea542f6be54b"} Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.375947 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5836f266843a9c18248bacd679bf75921ccbdd00f3b48a8df366ea542f6be54b" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.376456 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.487036 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q"] Nov 24 07:32:59 crc kubenswrapper[4809]: E1124 07:32:59.488115 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5f0f244-c1f0-4c1c-84ef-9c994b349739" containerName="collect-profiles" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.488311 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5f0f244-c1f0-4c1c-84ef-9c994b349739" containerName="collect-profiles" Nov 24 07:32:59 crc kubenswrapper[4809]: E1124 07:32:59.488449 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd7588f7-f400-41e6-a14e-d01e548130a3" containerName="registry-server" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.488561 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd7588f7-f400-41e6-a14e-d01e548130a3" containerName="registry-server" Nov 24 07:32:59 crc kubenswrapper[4809]: E1124 07:32:59.488685 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd7588f7-f400-41e6-a14e-d01e548130a3" containerName="extract-content" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.488812 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd7588f7-f400-41e6-a14e-d01e548130a3" containerName="extract-content" Nov 24 07:32:59 crc kubenswrapper[4809]: E1124 07:32:59.488949 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd55a20b-bdd2-40bf-85d7-c0b485352b8f" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.489283 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd55a20b-bdd2-40bf-85d7-c0b485352b8f" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 07:32:59 crc kubenswrapper[4809]: E1124 07:32:59.489430 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd7588f7-f400-41e6-a14e-d01e548130a3" containerName="extract-utilities" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.489774 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd7588f7-f400-41e6-a14e-d01e548130a3" containerName="extract-utilities" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.490324 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd7588f7-f400-41e6-a14e-d01e548130a3" containerName="registry-server" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.490806 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd55a20b-bdd2-40bf-85d7-c0b485352b8f" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.491043 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5f0f244-c1f0-4c1c-84ef-9c994b349739" containerName="collect-profiles" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.492266 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.497564 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.497742 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.498111 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.498461 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.498557 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.498741 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lx7ll" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.498466 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.505431 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q"] Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.612359 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.612649 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.612781 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.612915 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.613121 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.613249 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.613431 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwlfc\" (UniqueName: \"kubernetes.io/projected/4b9c2d10-ec36-41c7-985c-44ee750eddd0-kube-api-access-xwlfc\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.613615 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.613804 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.715624 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.715741 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.716132 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.716178 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.716206 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.716247 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.716291 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.716322 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.716430 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwlfc\" (UniqueName: \"kubernetes.io/projected/4b9c2d10-ec36-41c7-985c-44ee750eddd0-kube-api-access-xwlfc\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.717517 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.719182 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.722127 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.722182 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.724953 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.729351 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.729389 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.729391 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.737718 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwlfc\" (UniqueName: \"kubernetes.io/projected/4b9c2d10-ec36-41c7-985c-44ee750eddd0-kube-api-access-xwlfc\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wsk2q\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:32:59 crc kubenswrapper[4809]: I1124 07:32:59.823849 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:33:00 crc kubenswrapper[4809]: I1124 07:33:00.143879 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q"] Nov 24 07:33:00 crc kubenswrapper[4809]: I1124 07:33:00.154158 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 07:33:00 crc kubenswrapper[4809]: I1124 07:33:00.389941 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" event={"ID":"4b9c2d10-ec36-41c7-985c-44ee750eddd0","Type":"ContainerStarted","Data":"ce4ca73578ba614513bafb1e1eb58a92eb63615f1a25366b92a12dc00e9252a4"} Nov 24 07:33:01 crc kubenswrapper[4809]: I1124 07:33:01.401117 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" event={"ID":"4b9c2d10-ec36-41c7-985c-44ee750eddd0","Type":"ContainerStarted","Data":"d70a270a7334692e06e39592d2c0d94023ce85fbf86ffdb44c8fc6105856ad11"} Nov 24 07:33:01 crc kubenswrapper[4809]: I1124 07:33:01.891511 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:33:01 crc kubenswrapper[4809]: E1124 07:33:01.892014 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:33:15 crc kubenswrapper[4809]: I1124 07:33:15.891500 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:33:15 crc kubenswrapper[4809]: E1124 07:33:15.892293 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:33:26 crc kubenswrapper[4809]: I1124 07:33:26.892417 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:33:26 crc kubenswrapper[4809]: E1124 07:33:26.893631 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:33:41 crc kubenswrapper[4809]: I1124 07:33:41.892355 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:33:41 crc kubenswrapper[4809]: E1124 07:33:41.893647 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:33:52 crc kubenswrapper[4809]: I1124 07:33:52.892014 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:33:52 crc kubenswrapper[4809]: E1124 07:33:52.893101 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:34:04 crc kubenswrapper[4809]: I1124 07:34:04.891136 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:34:04 crc kubenswrapper[4809]: E1124 07:34:04.892171 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:34:19 crc kubenswrapper[4809]: I1124 07:34:19.891598 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:34:19 crc kubenswrapper[4809]: E1124 07:34:19.892662 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:34:33 crc kubenswrapper[4809]: I1124 07:34:33.892214 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:34:33 crc kubenswrapper[4809]: E1124 07:34:33.893509 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:34:47 crc kubenswrapper[4809]: I1124 07:34:47.891996 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:34:47 crc kubenswrapper[4809]: E1124 07:34:47.892945 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:35:01 crc kubenswrapper[4809]: I1124 07:35:01.891430 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:35:01 crc kubenswrapper[4809]: E1124 07:35:01.892068 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:35:12 crc kubenswrapper[4809]: I1124 07:35:12.891338 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:35:12 crc kubenswrapper[4809]: E1124 07:35:12.892025 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:35:23 crc kubenswrapper[4809]: I1124 07:35:23.890766 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:35:23 crc kubenswrapper[4809]: E1124 07:35:23.891722 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:35:35 crc kubenswrapper[4809]: I1124 07:35:35.890831 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:35:35 crc kubenswrapper[4809]: E1124 07:35:35.891567 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:35:50 crc kubenswrapper[4809]: I1124 07:35:50.896236 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:35:50 crc kubenswrapper[4809]: E1124 07:35:50.897595 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:36:03 crc kubenswrapper[4809]: I1124 07:36:03.890668 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:36:03 crc kubenswrapper[4809]: E1124 07:36:03.891332 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:36:09 crc kubenswrapper[4809]: I1124 07:36:09.243090 4809 generic.go:334] "Generic (PLEG): container finished" podID="4b9c2d10-ec36-41c7-985c-44ee750eddd0" containerID="d70a270a7334692e06e39592d2c0d94023ce85fbf86ffdb44c8fc6105856ad11" exitCode=0 Nov 24 07:36:09 crc kubenswrapper[4809]: I1124 07:36:09.243133 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" event={"ID":"4b9c2d10-ec36-41c7-985c-44ee750eddd0","Type":"ContainerDied","Data":"d70a270a7334692e06e39592d2c0d94023ce85fbf86ffdb44c8fc6105856ad11"} Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.719416 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.884747 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-ssh-key\") pod \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.884849 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-combined-ca-bundle\") pod \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.884877 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-cell1-compute-config-1\") pod \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.885186 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-migration-ssh-key-0\") pod \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.885249 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-inventory\") pod \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.885340 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-cell1-compute-config-0\") pod \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.885368 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-migration-ssh-key-1\") pod \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.885402 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-extra-config-0\") pod \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.885465 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwlfc\" (UniqueName: \"kubernetes.io/projected/4b9c2d10-ec36-41c7-985c-44ee750eddd0-kube-api-access-xwlfc\") pod \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\" (UID: \"4b9c2d10-ec36-41c7-985c-44ee750eddd0\") " Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.891469 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b9c2d10-ec36-41c7-985c-44ee750eddd0-kube-api-access-xwlfc" (OuterVolumeSpecName: "kube-api-access-xwlfc") pod "4b9c2d10-ec36-41c7-985c-44ee750eddd0" (UID: "4b9c2d10-ec36-41c7-985c-44ee750eddd0"). InnerVolumeSpecName "kube-api-access-xwlfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.893419 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "4b9c2d10-ec36-41c7-985c-44ee750eddd0" (UID: "4b9c2d10-ec36-41c7-985c-44ee750eddd0"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.914928 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "4b9c2d10-ec36-41c7-985c-44ee750eddd0" (UID: "4b9c2d10-ec36-41c7-985c-44ee750eddd0"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.915670 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-inventory" (OuterVolumeSpecName: "inventory") pod "4b9c2d10-ec36-41c7-985c-44ee750eddd0" (UID: "4b9c2d10-ec36-41c7-985c-44ee750eddd0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.917355 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "4b9c2d10-ec36-41c7-985c-44ee750eddd0" (UID: "4b9c2d10-ec36-41c7-985c-44ee750eddd0"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.923629 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "4b9c2d10-ec36-41c7-985c-44ee750eddd0" (UID: "4b9c2d10-ec36-41c7-985c-44ee750eddd0"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.932504 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "4b9c2d10-ec36-41c7-985c-44ee750eddd0" (UID: "4b9c2d10-ec36-41c7-985c-44ee750eddd0"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.942021 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4b9c2d10-ec36-41c7-985c-44ee750eddd0" (UID: "4b9c2d10-ec36-41c7-985c-44ee750eddd0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.942164 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "4b9c2d10-ec36-41c7-985c-44ee750eddd0" (UID: "4b9c2d10-ec36-41c7-985c-44ee750eddd0"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.987984 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.988012 4809 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.988022 4809 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.988031 4809 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.988040 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwlfc\" (UniqueName: \"kubernetes.io/projected/4b9c2d10-ec36-41c7-985c-44ee750eddd0-kube-api-access-xwlfc\") on node \"crc\" DevicePath \"\"" Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.988047 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.988056 4809 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.988064 4809 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 24 07:36:10 crc kubenswrapper[4809]: I1124 07:36:10.988072 4809 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4b9c2d10-ec36-41c7-985c-44ee750eddd0-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.296878 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" event={"ID":"4b9c2d10-ec36-41c7-985c-44ee750eddd0","Type":"ContainerDied","Data":"ce4ca73578ba614513bafb1e1eb58a92eb63615f1a25366b92a12dc00e9252a4"} Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.296930 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce4ca73578ba614513bafb1e1eb58a92eb63615f1a25366b92a12dc00e9252a4" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.297016 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wsk2q" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.377226 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd"] Nov 24 07:36:11 crc kubenswrapper[4809]: E1124 07:36:11.377610 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b9c2d10-ec36-41c7-985c-44ee750eddd0" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.377628 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b9c2d10-ec36-41c7-985c-44ee750eddd0" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.377807 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b9c2d10-ec36-41c7-985c-44ee750eddd0" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.378469 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.380432 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.380796 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lx7ll" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.381527 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.381708 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.382569 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.390253 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd"] Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.500060 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.500198 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.500240 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.500301 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldrjr\" (UniqueName: \"kubernetes.io/projected/797b01f0-7757-495a-ad23-3973353832c1-kube-api-access-ldrjr\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.500493 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.500574 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.500618 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.602250 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.602291 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.602322 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.602642 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.602699 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.602729 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.602791 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldrjr\" (UniqueName: \"kubernetes.io/projected/797b01f0-7757-495a-ad23-3973353832c1-kube-api-access-ldrjr\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.606339 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.606354 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.606683 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.607607 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.609050 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.609072 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.624627 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldrjr\" (UniqueName: \"kubernetes.io/projected/797b01f0-7757-495a-ad23-3973353832c1-kube-api-access-ldrjr\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:36:11 crc kubenswrapper[4809]: I1124 07:36:11.694644 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:36:12 crc kubenswrapper[4809]: I1124 07:36:12.222602 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd"] Nov 24 07:36:12 crc kubenswrapper[4809]: I1124 07:36:12.306091 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" event={"ID":"797b01f0-7757-495a-ad23-3973353832c1","Type":"ContainerStarted","Data":"35a123e293432368e60f92803ff291839d73079666b31d6fab353d9e762e2029"} Nov 24 07:36:13 crc kubenswrapper[4809]: I1124 07:36:13.331624 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" event={"ID":"797b01f0-7757-495a-ad23-3973353832c1","Type":"ContainerStarted","Data":"656d8e88297e824457410ee4b945ad24f136b2cca997b49c65f6e68ab2c8478f"} Nov 24 07:36:13 crc kubenswrapper[4809]: I1124 07:36:13.352319 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" podStartSLOduration=1.657920573 podStartE2EDuration="2.352296385s" podCreationTimestamp="2025-11-24 07:36:11 +0000 UTC" firstStartedPulling="2025-11-24 07:36:12.224267241 +0000 UTC m=+2472.124858946" lastFinishedPulling="2025-11-24 07:36:12.918643013 +0000 UTC m=+2472.819234758" observedRunningTime="2025-11-24 07:36:13.349640894 +0000 UTC m=+2473.250232609" watchObservedRunningTime="2025-11-24 07:36:13.352296385 +0000 UTC m=+2473.252888100" Nov 24 07:36:17 crc kubenswrapper[4809]: I1124 07:36:17.891377 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:36:17 crc kubenswrapper[4809]: E1124 07:36:17.892246 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:36:29 crc kubenswrapper[4809]: I1124 07:36:29.890844 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:36:30 crc kubenswrapper[4809]: I1124 07:36:30.510263 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerStarted","Data":"630f55bb399f3bfd04e8309b3cf29adad95cffa6451d4ba31d28cc6a268df881"} Nov 24 07:38:48 crc kubenswrapper[4809]: I1124 07:38:48.043797 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:38:48 crc kubenswrapper[4809]: I1124 07:38:48.044458 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:38:48 crc kubenswrapper[4809]: I1124 07:38:48.827439 4809 generic.go:334] "Generic (PLEG): container finished" podID="797b01f0-7757-495a-ad23-3973353832c1" containerID="656d8e88297e824457410ee4b945ad24f136b2cca997b49c65f6e68ab2c8478f" exitCode=0 Nov 24 07:38:48 crc kubenswrapper[4809]: I1124 07:38:48.827485 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" event={"ID":"797b01f0-7757-495a-ad23-3973353832c1","Type":"ContainerDied","Data":"656d8e88297e824457410ee4b945ad24f136b2cca997b49c65f6e68ab2c8478f"} Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.244300 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.300842 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldrjr\" (UniqueName: \"kubernetes.io/projected/797b01f0-7757-495a-ad23-3973353832c1-kube-api-access-ldrjr\") pod \"797b01f0-7757-495a-ad23-3973353832c1\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.300941 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ceilometer-compute-config-data-2\") pod \"797b01f0-7757-495a-ad23-3973353832c1\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.301047 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-telemetry-combined-ca-bundle\") pod \"797b01f0-7757-495a-ad23-3973353832c1\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.301107 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ssh-key\") pod \"797b01f0-7757-495a-ad23-3973353832c1\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.301243 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ceilometer-compute-config-data-1\") pod \"797b01f0-7757-495a-ad23-3973353832c1\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.301356 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-inventory\") pod \"797b01f0-7757-495a-ad23-3973353832c1\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.301430 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ceilometer-compute-config-data-0\") pod \"797b01f0-7757-495a-ad23-3973353832c1\" (UID: \"797b01f0-7757-495a-ad23-3973353832c1\") " Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.310260 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "797b01f0-7757-495a-ad23-3973353832c1" (UID: "797b01f0-7757-495a-ad23-3973353832c1"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.314854 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/797b01f0-7757-495a-ad23-3973353832c1-kube-api-access-ldrjr" (OuterVolumeSpecName: "kube-api-access-ldrjr") pod "797b01f0-7757-495a-ad23-3973353832c1" (UID: "797b01f0-7757-495a-ad23-3973353832c1"). InnerVolumeSpecName "kube-api-access-ldrjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.333187 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "797b01f0-7757-495a-ad23-3973353832c1" (UID: "797b01f0-7757-495a-ad23-3973353832c1"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.335818 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-inventory" (OuterVolumeSpecName: "inventory") pod "797b01f0-7757-495a-ad23-3973353832c1" (UID: "797b01f0-7757-495a-ad23-3973353832c1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.337880 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "797b01f0-7757-495a-ad23-3973353832c1" (UID: "797b01f0-7757-495a-ad23-3973353832c1"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.346325 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "797b01f0-7757-495a-ad23-3973353832c1" (UID: "797b01f0-7757-495a-ad23-3973353832c1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.367446 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "797b01f0-7757-495a-ad23-3973353832c1" (UID: "797b01f0-7757-495a-ad23-3973353832c1"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.404167 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.404195 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.404208 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldrjr\" (UniqueName: \"kubernetes.io/projected/797b01f0-7757-495a-ad23-3973353832c1-kube-api-access-ldrjr\") on node \"crc\" DevicePath \"\"" Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.404220 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.404229 4809 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.404237 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.404247 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/797b01f0-7757-495a-ad23-3973353832c1-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.848784 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" event={"ID":"797b01f0-7757-495a-ad23-3973353832c1","Type":"ContainerDied","Data":"35a123e293432368e60f92803ff291839d73079666b31d6fab353d9e762e2029"} Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.849192 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35a123e293432368e60f92803ff291839d73079666b31d6fab353d9e762e2029" Nov 24 07:38:50 crc kubenswrapper[4809]: I1124 07:38:50.848989 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd" Nov 24 07:38:51 crc kubenswrapper[4809]: E1124 07:38:51.123994 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod797b01f0_7757_495a_ad23_3973353832c1.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod797b01f0_7757_495a_ad23_3973353832c1.slice/crio-35a123e293432368e60f92803ff291839d73079666b31d6fab353d9e762e2029\": RecentStats: unable to find data in memory cache]" Nov 24 07:38:57 crc kubenswrapper[4809]: I1124 07:38:57.531804 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ws6wx"] Nov 24 07:38:57 crc kubenswrapper[4809]: E1124 07:38:57.532751 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="797b01f0-7757-495a-ad23-3973353832c1" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 07:38:57 crc kubenswrapper[4809]: I1124 07:38:57.532768 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="797b01f0-7757-495a-ad23-3973353832c1" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 07:38:57 crc kubenswrapper[4809]: I1124 07:38:57.533012 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="797b01f0-7757-495a-ad23-3973353832c1" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 07:38:57 crc kubenswrapper[4809]: I1124 07:38:57.534287 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ws6wx" Nov 24 07:38:57 crc kubenswrapper[4809]: I1124 07:38:57.544340 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ws6wx"] Nov 24 07:38:57 crc kubenswrapper[4809]: I1124 07:38:57.654440 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfrbv\" (UniqueName: \"kubernetes.io/projected/422d766e-1186-4b50-b09c-2fcced744b5d-kube-api-access-gfrbv\") pod \"community-operators-ws6wx\" (UID: \"422d766e-1186-4b50-b09c-2fcced744b5d\") " pod="openshift-marketplace/community-operators-ws6wx" Nov 24 07:38:57 crc kubenswrapper[4809]: I1124 07:38:57.654500 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/422d766e-1186-4b50-b09c-2fcced744b5d-catalog-content\") pod \"community-operators-ws6wx\" (UID: \"422d766e-1186-4b50-b09c-2fcced744b5d\") " pod="openshift-marketplace/community-operators-ws6wx" Nov 24 07:38:57 crc kubenswrapper[4809]: I1124 07:38:57.654537 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/422d766e-1186-4b50-b09c-2fcced744b5d-utilities\") pod \"community-operators-ws6wx\" (UID: \"422d766e-1186-4b50-b09c-2fcced744b5d\") " pod="openshift-marketplace/community-operators-ws6wx" Nov 24 07:38:57 crc kubenswrapper[4809]: I1124 07:38:57.756710 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/422d766e-1186-4b50-b09c-2fcced744b5d-utilities\") pod \"community-operators-ws6wx\" (UID: \"422d766e-1186-4b50-b09c-2fcced744b5d\") " pod="openshift-marketplace/community-operators-ws6wx" Nov 24 07:38:57 crc kubenswrapper[4809]: I1124 07:38:57.756895 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfrbv\" (UniqueName: \"kubernetes.io/projected/422d766e-1186-4b50-b09c-2fcced744b5d-kube-api-access-gfrbv\") pod \"community-operators-ws6wx\" (UID: \"422d766e-1186-4b50-b09c-2fcced744b5d\") " pod="openshift-marketplace/community-operators-ws6wx" Nov 24 07:38:57 crc kubenswrapper[4809]: I1124 07:38:57.756935 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/422d766e-1186-4b50-b09c-2fcced744b5d-catalog-content\") pod \"community-operators-ws6wx\" (UID: \"422d766e-1186-4b50-b09c-2fcced744b5d\") " pod="openshift-marketplace/community-operators-ws6wx" Nov 24 07:38:57 crc kubenswrapper[4809]: I1124 07:38:57.757382 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/422d766e-1186-4b50-b09c-2fcced744b5d-utilities\") pod \"community-operators-ws6wx\" (UID: \"422d766e-1186-4b50-b09c-2fcced744b5d\") " pod="openshift-marketplace/community-operators-ws6wx" Nov 24 07:38:57 crc kubenswrapper[4809]: I1124 07:38:57.757410 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/422d766e-1186-4b50-b09c-2fcced744b5d-catalog-content\") pod \"community-operators-ws6wx\" (UID: \"422d766e-1186-4b50-b09c-2fcced744b5d\") " pod="openshift-marketplace/community-operators-ws6wx" Nov 24 07:38:57 crc kubenswrapper[4809]: I1124 07:38:57.785611 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfrbv\" (UniqueName: \"kubernetes.io/projected/422d766e-1186-4b50-b09c-2fcced744b5d-kube-api-access-gfrbv\") pod \"community-operators-ws6wx\" (UID: \"422d766e-1186-4b50-b09c-2fcced744b5d\") " pod="openshift-marketplace/community-operators-ws6wx" Nov 24 07:38:57 crc kubenswrapper[4809]: I1124 07:38:57.869317 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ws6wx" Nov 24 07:38:58 crc kubenswrapper[4809]: I1124 07:38:58.466786 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ws6wx"] Nov 24 07:38:58 crc kubenswrapper[4809]: I1124 07:38:58.954612 4809 generic.go:334] "Generic (PLEG): container finished" podID="422d766e-1186-4b50-b09c-2fcced744b5d" containerID="5132c5886dd964149c5db94c6675120ad297538a2ff91d544d28112d64f77e90" exitCode=0 Nov 24 07:38:58 crc kubenswrapper[4809]: I1124 07:38:58.954715 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ws6wx" event={"ID":"422d766e-1186-4b50-b09c-2fcced744b5d","Type":"ContainerDied","Data":"5132c5886dd964149c5db94c6675120ad297538a2ff91d544d28112d64f77e90"} Nov 24 07:38:58 crc kubenswrapper[4809]: I1124 07:38:58.954923 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ws6wx" event={"ID":"422d766e-1186-4b50-b09c-2fcced744b5d","Type":"ContainerStarted","Data":"aa82908f00278097034f7050801d27aa3a0dfcaa0e82bf3385ff71d77d9e0ff3"} Nov 24 07:38:58 crc kubenswrapper[4809]: I1124 07:38:58.956795 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 07:39:03 crc kubenswrapper[4809]: I1124 07:39:03.010134 4809 generic.go:334] "Generic (PLEG): container finished" podID="422d766e-1186-4b50-b09c-2fcced744b5d" containerID="6c9e606568e5ad8c80cb943201284c442c128edd3015bc741b453545fb6dff5c" exitCode=0 Nov 24 07:39:03 crc kubenswrapper[4809]: I1124 07:39:03.010226 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ws6wx" event={"ID":"422d766e-1186-4b50-b09c-2fcced744b5d","Type":"ContainerDied","Data":"6c9e606568e5ad8c80cb943201284c442c128edd3015bc741b453545fb6dff5c"} Nov 24 07:39:04 crc kubenswrapper[4809]: I1124 07:39:04.022221 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ws6wx" event={"ID":"422d766e-1186-4b50-b09c-2fcced744b5d","Type":"ContainerStarted","Data":"c7c64fe755d3c194409c2edfe7328bb8c2c5d1e1c41b0ee642f37eaf4c0b43dd"} Nov 24 07:39:04 crc kubenswrapper[4809]: I1124 07:39:04.052895 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ws6wx" podStartSLOduration=2.347764822 podStartE2EDuration="7.052875218s" podCreationTimestamp="2025-11-24 07:38:57 +0000 UTC" firstStartedPulling="2025-11-24 07:38:58.956581157 +0000 UTC m=+2638.857172862" lastFinishedPulling="2025-11-24 07:39:03.661691553 +0000 UTC m=+2643.562283258" observedRunningTime="2025-11-24 07:39:04.040120719 +0000 UTC m=+2643.940712444" watchObservedRunningTime="2025-11-24 07:39:04.052875218 +0000 UTC m=+2643.953466943" Nov 24 07:39:07 crc kubenswrapper[4809]: I1124 07:39:07.869560 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ws6wx" Nov 24 07:39:07 crc kubenswrapper[4809]: I1124 07:39:07.869834 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ws6wx" Nov 24 07:39:07 crc kubenswrapper[4809]: I1124 07:39:07.932376 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ws6wx" Nov 24 07:39:08 crc kubenswrapper[4809]: I1124 07:39:08.117621 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ws6wx" Nov 24 07:39:08 crc kubenswrapper[4809]: I1124 07:39:08.231561 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ws6wx"] Nov 24 07:39:08 crc kubenswrapper[4809]: I1124 07:39:08.275718 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rhg7d"] Nov 24 07:39:08 crc kubenswrapper[4809]: I1124 07:39:08.275986 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rhg7d" podUID="41964283-35a3-4b8d-8e54-c3c7518f62ee" containerName="registry-server" containerID="cri-o://ff3d073588c45d12604f414b860d30c70c95f29b7bfc326e4990a89750ef5d91" gracePeriod=2 Nov 24 07:39:08 crc kubenswrapper[4809]: I1124 07:39:08.807485 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rhg7d" Nov 24 07:39:08 crc kubenswrapper[4809]: I1124 07:39:08.895511 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nnb5\" (UniqueName: \"kubernetes.io/projected/41964283-35a3-4b8d-8e54-c3c7518f62ee-kube-api-access-6nnb5\") pod \"41964283-35a3-4b8d-8e54-c3c7518f62ee\" (UID: \"41964283-35a3-4b8d-8e54-c3c7518f62ee\") " Nov 24 07:39:08 crc kubenswrapper[4809]: I1124 07:39:08.895668 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41964283-35a3-4b8d-8e54-c3c7518f62ee-catalog-content\") pod \"41964283-35a3-4b8d-8e54-c3c7518f62ee\" (UID: \"41964283-35a3-4b8d-8e54-c3c7518f62ee\") " Nov 24 07:39:08 crc kubenswrapper[4809]: I1124 07:39:08.899281 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41964283-35a3-4b8d-8e54-c3c7518f62ee-utilities\") pod \"41964283-35a3-4b8d-8e54-c3c7518f62ee\" (UID: \"41964283-35a3-4b8d-8e54-c3c7518f62ee\") " Nov 24 07:39:08 crc kubenswrapper[4809]: I1124 07:39:08.901382 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41964283-35a3-4b8d-8e54-c3c7518f62ee-kube-api-access-6nnb5" (OuterVolumeSpecName: "kube-api-access-6nnb5") pod "41964283-35a3-4b8d-8e54-c3c7518f62ee" (UID: "41964283-35a3-4b8d-8e54-c3c7518f62ee"). InnerVolumeSpecName "kube-api-access-6nnb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:39:08 crc kubenswrapper[4809]: I1124 07:39:08.901693 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41964283-35a3-4b8d-8e54-c3c7518f62ee-utilities" (OuterVolumeSpecName: "utilities") pod "41964283-35a3-4b8d-8e54-c3c7518f62ee" (UID: "41964283-35a3-4b8d-8e54-c3c7518f62ee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:39:08 crc kubenswrapper[4809]: I1124 07:39:08.952613 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41964283-35a3-4b8d-8e54-c3c7518f62ee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "41964283-35a3-4b8d-8e54-c3c7518f62ee" (UID: "41964283-35a3-4b8d-8e54-c3c7518f62ee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:39:09 crc kubenswrapper[4809]: I1124 07:39:09.002534 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nnb5\" (UniqueName: \"kubernetes.io/projected/41964283-35a3-4b8d-8e54-c3c7518f62ee-kube-api-access-6nnb5\") on node \"crc\" DevicePath \"\"" Nov 24 07:39:09 crc kubenswrapper[4809]: I1124 07:39:09.002562 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41964283-35a3-4b8d-8e54-c3c7518f62ee-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:39:09 crc kubenswrapper[4809]: I1124 07:39:09.002573 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41964283-35a3-4b8d-8e54-c3c7518f62ee-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:39:09 crc kubenswrapper[4809]: I1124 07:39:09.065386 4809 generic.go:334] "Generic (PLEG): container finished" podID="41964283-35a3-4b8d-8e54-c3c7518f62ee" containerID="ff3d073588c45d12604f414b860d30c70c95f29b7bfc326e4990a89750ef5d91" exitCode=0 Nov 24 07:39:09 crc kubenswrapper[4809]: I1124 07:39:09.065512 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rhg7d" Nov 24 07:39:09 crc kubenswrapper[4809]: I1124 07:39:09.065564 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rhg7d" event={"ID":"41964283-35a3-4b8d-8e54-c3c7518f62ee","Type":"ContainerDied","Data":"ff3d073588c45d12604f414b860d30c70c95f29b7bfc326e4990a89750ef5d91"} Nov 24 07:39:09 crc kubenswrapper[4809]: I1124 07:39:09.065596 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rhg7d" event={"ID":"41964283-35a3-4b8d-8e54-c3c7518f62ee","Type":"ContainerDied","Data":"164aca2d37836075cf656b2386a2d8d275297ccb2a1b86baef840d7e2cfc09ee"} Nov 24 07:39:09 crc kubenswrapper[4809]: I1124 07:39:09.065614 4809 scope.go:117] "RemoveContainer" containerID="ff3d073588c45d12604f414b860d30c70c95f29b7bfc326e4990a89750ef5d91" Nov 24 07:39:09 crc kubenswrapper[4809]: I1124 07:39:09.093855 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rhg7d"] Nov 24 07:39:09 crc kubenswrapper[4809]: I1124 07:39:09.102117 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rhg7d"] Nov 24 07:39:09 crc kubenswrapper[4809]: I1124 07:39:09.131437 4809 scope.go:117] "RemoveContainer" containerID="15bb3bd9449fd67416a8a8e56167b5b38291595472c72e75431227d385a2bf27" Nov 24 07:39:09 crc kubenswrapper[4809]: I1124 07:39:09.161572 4809 scope.go:117] "RemoveContainer" containerID="03130db63b837c63ae279deb3214140c028d66ff03bbd571f524d761c68f8148" Nov 24 07:39:09 crc kubenswrapper[4809]: I1124 07:39:09.238142 4809 scope.go:117] "RemoveContainer" containerID="ff3d073588c45d12604f414b860d30c70c95f29b7bfc326e4990a89750ef5d91" Nov 24 07:39:09 crc kubenswrapper[4809]: E1124 07:39:09.238708 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff3d073588c45d12604f414b860d30c70c95f29b7bfc326e4990a89750ef5d91\": container with ID starting with ff3d073588c45d12604f414b860d30c70c95f29b7bfc326e4990a89750ef5d91 not found: ID does not exist" containerID="ff3d073588c45d12604f414b860d30c70c95f29b7bfc326e4990a89750ef5d91" Nov 24 07:39:09 crc kubenswrapper[4809]: I1124 07:39:09.238749 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff3d073588c45d12604f414b860d30c70c95f29b7bfc326e4990a89750ef5d91"} err="failed to get container status \"ff3d073588c45d12604f414b860d30c70c95f29b7bfc326e4990a89750ef5d91\": rpc error: code = NotFound desc = could not find container \"ff3d073588c45d12604f414b860d30c70c95f29b7bfc326e4990a89750ef5d91\": container with ID starting with ff3d073588c45d12604f414b860d30c70c95f29b7bfc326e4990a89750ef5d91 not found: ID does not exist" Nov 24 07:39:09 crc kubenswrapper[4809]: I1124 07:39:09.238775 4809 scope.go:117] "RemoveContainer" containerID="15bb3bd9449fd67416a8a8e56167b5b38291595472c72e75431227d385a2bf27" Nov 24 07:39:09 crc kubenswrapper[4809]: E1124 07:39:09.239348 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15bb3bd9449fd67416a8a8e56167b5b38291595472c72e75431227d385a2bf27\": container with ID starting with 15bb3bd9449fd67416a8a8e56167b5b38291595472c72e75431227d385a2bf27 not found: ID does not exist" containerID="15bb3bd9449fd67416a8a8e56167b5b38291595472c72e75431227d385a2bf27" Nov 24 07:39:09 crc kubenswrapper[4809]: I1124 07:39:09.239400 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15bb3bd9449fd67416a8a8e56167b5b38291595472c72e75431227d385a2bf27"} err="failed to get container status \"15bb3bd9449fd67416a8a8e56167b5b38291595472c72e75431227d385a2bf27\": rpc error: code = NotFound desc = could not find container \"15bb3bd9449fd67416a8a8e56167b5b38291595472c72e75431227d385a2bf27\": container with ID starting with 15bb3bd9449fd67416a8a8e56167b5b38291595472c72e75431227d385a2bf27 not found: ID does not exist" Nov 24 07:39:09 crc kubenswrapper[4809]: I1124 07:39:09.239434 4809 scope.go:117] "RemoveContainer" containerID="03130db63b837c63ae279deb3214140c028d66ff03bbd571f524d761c68f8148" Nov 24 07:39:09 crc kubenswrapper[4809]: E1124 07:39:09.239780 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03130db63b837c63ae279deb3214140c028d66ff03bbd571f524d761c68f8148\": container with ID starting with 03130db63b837c63ae279deb3214140c028d66ff03bbd571f524d761c68f8148 not found: ID does not exist" containerID="03130db63b837c63ae279deb3214140c028d66ff03bbd571f524d761c68f8148" Nov 24 07:39:09 crc kubenswrapper[4809]: I1124 07:39:09.239807 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03130db63b837c63ae279deb3214140c028d66ff03bbd571f524d761c68f8148"} err="failed to get container status \"03130db63b837c63ae279deb3214140c028d66ff03bbd571f524d761c68f8148\": rpc error: code = NotFound desc = could not find container \"03130db63b837c63ae279deb3214140c028d66ff03bbd571f524d761c68f8148\": container with ID starting with 03130db63b837c63ae279deb3214140c028d66ff03bbd571f524d761c68f8148 not found: ID does not exist" Nov 24 07:39:10 crc kubenswrapper[4809]: I1124 07:39:10.909137 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41964283-35a3-4b8d-8e54-c3c7518f62ee" path="/var/lib/kubelet/pods/41964283-35a3-4b8d-8e54-c3c7518f62ee/volumes" Nov 24 07:39:18 crc kubenswrapper[4809]: I1124 07:39:18.043726 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:39:18 crc kubenswrapper[4809]: I1124 07:39:18.044329 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:39:37 crc kubenswrapper[4809]: I1124 07:39:37.653351 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nzgcn"] Nov 24 07:39:37 crc kubenswrapper[4809]: E1124 07:39:37.654564 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41964283-35a3-4b8d-8e54-c3c7518f62ee" containerName="registry-server" Nov 24 07:39:37 crc kubenswrapper[4809]: I1124 07:39:37.654583 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="41964283-35a3-4b8d-8e54-c3c7518f62ee" containerName="registry-server" Nov 24 07:39:37 crc kubenswrapper[4809]: E1124 07:39:37.654608 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41964283-35a3-4b8d-8e54-c3c7518f62ee" containerName="extract-utilities" Nov 24 07:39:37 crc kubenswrapper[4809]: I1124 07:39:37.654617 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="41964283-35a3-4b8d-8e54-c3c7518f62ee" containerName="extract-utilities" Nov 24 07:39:37 crc kubenswrapper[4809]: E1124 07:39:37.654641 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41964283-35a3-4b8d-8e54-c3c7518f62ee" containerName="extract-content" Nov 24 07:39:37 crc kubenswrapper[4809]: I1124 07:39:37.654649 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="41964283-35a3-4b8d-8e54-c3c7518f62ee" containerName="extract-content" Nov 24 07:39:37 crc kubenswrapper[4809]: I1124 07:39:37.654884 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="41964283-35a3-4b8d-8e54-c3c7518f62ee" containerName="registry-server" Nov 24 07:39:37 crc kubenswrapper[4809]: I1124 07:39:37.656637 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nzgcn" Nov 24 07:39:37 crc kubenswrapper[4809]: I1124 07:39:37.678729 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nzgcn"] Nov 24 07:39:37 crc kubenswrapper[4809]: I1124 07:39:37.773491 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17720a8a-d721-431b-8bc4-1efd5ac1a634-utilities\") pod \"redhat-operators-nzgcn\" (UID: \"17720a8a-d721-431b-8bc4-1efd5ac1a634\") " pod="openshift-marketplace/redhat-operators-nzgcn" Nov 24 07:39:37 crc kubenswrapper[4809]: I1124 07:39:37.773638 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbmff\" (UniqueName: \"kubernetes.io/projected/17720a8a-d721-431b-8bc4-1efd5ac1a634-kube-api-access-nbmff\") pod \"redhat-operators-nzgcn\" (UID: \"17720a8a-d721-431b-8bc4-1efd5ac1a634\") " pod="openshift-marketplace/redhat-operators-nzgcn" Nov 24 07:39:37 crc kubenswrapper[4809]: I1124 07:39:37.773757 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17720a8a-d721-431b-8bc4-1efd5ac1a634-catalog-content\") pod \"redhat-operators-nzgcn\" (UID: \"17720a8a-d721-431b-8bc4-1efd5ac1a634\") " pod="openshift-marketplace/redhat-operators-nzgcn" Nov 24 07:39:37 crc kubenswrapper[4809]: I1124 07:39:37.875348 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbmff\" (UniqueName: \"kubernetes.io/projected/17720a8a-d721-431b-8bc4-1efd5ac1a634-kube-api-access-nbmff\") pod \"redhat-operators-nzgcn\" (UID: \"17720a8a-d721-431b-8bc4-1efd5ac1a634\") " pod="openshift-marketplace/redhat-operators-nzgcn" Nov 24 07:39:37 crc kubenswrapper[4809]: I1124 07:39:37.875457 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17720a8a-d721-431b-8bc4-1efd5ac1a634-catalog-content\") pod \"redhat-operators-nzgcn\" (UID: \"17720a8a-d721-431b-8bc4-1efd5ac1a634\") " pod="openshift-marketplace/redhat-operators-nzgcn" Nov 24 07:39:37 crc kubenswrapper[4809]: I1124 07:39:37.875510 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17720a8a-d721-431b-8bc4-1efd5ac1a634-utilities\") pod \"redhat-operators-nzgcn\" (UID: \"17720a8a-d721-431b-8bc4-1efd5ac1a634\") " pod="openshift-marketplace/redhat-operators-nzgcn" Nov 24 07:39:37 crc kubenswrapper[4809]: I1124 07:39:37.875958 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17720a8a-d721-431b-8bc4-1efd5ac1a634-utilities\") pod \"redhat-operators-nzgcn\" (UID: \"17720a8a-d721-431b-8bc4-1efd5ac1a634\") " pod="openshift-marketplace/redhat-operators-nzgcn" Nov 24 07:39:37 crc kubenswrapper[4809]: I1124 07:39:37.876109 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17720a8a-d721-431b-8bc4-1efd5ac1a634-catalog-content\") pod \"redhat-operators-nzgcn\" (UID: \"17720a8a-d721-431b-8bc4-1efd5ac1a634\") " pod="openshift-marketplace/redhat-operators-nzgcn" Nov 24 07:39:37 crc kubenswrapper[4809]: I1124 07:39:37.914189 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbmff\" (UniqueName: \"kubernetes.io/projected/17720a8a-d721-431b-8bc4-1efd5ac1a634-kube-api-access-nbmff\") pod \"redhat-operators-nzgcn\" (UID: \"17720a8a-d721-431b-8bc4-1efd5ac1a634\") " pod="openshift-marketplace/redhat-operators-nzgcn" Nov 24 07:39:37 crc kubenswrapper[4809]: I1124 07:39:37.987325 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nzgcn" Nov 24 07:39:38 crc kubenswrapper[4809]: I1124 07:39:38.486502 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nzgcn"] Nov 24 07:39:38 crc kubenswrapper[4809]: I1124 07:39:38.998925 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wxg6r"] Nov 24 07:39:39 crc kubenswrapper[4809]: I1124 07:39:39.002178 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wxg6r" Nov 24 07:39:39 crc kubenswrapper[4809]: I1124 07:39:39.015763 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wxg6r"] Nov 24 07:39:39 crc kubenswrapper[4809]: I1124 07:39:39.097302 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-555xs\" (UniqueName: \"kubernetes.io/projected/3bf14564-2c89-4c30-abc1-06bddc04a0d6-kube-api-access-555xs\") pod \"certified-operators-wxg6r\" (UID: \"3bf14564-2c89-4c30-abc1-06bddc04a0d6\") " pod="openshift-marketplace/certified-operators-wxg6r" Nov 24 07:39:39 crc kubenswrapper[4809]: I1124 07:39:39.097514 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bf14564-2c89-4c30-abc1-06bddc04a0d6-catalog-content\") pod \"certified-operators-wxg6r\" (UID: \"3bf14564-2c89-4c30-abc1-06bddc04a0d6\") " pod="openshift-marketplace/certified-operators-wxg6r" Nov 24 07:39:39 crc kubenswrapper[4809]: I1124 07:39:39.097562 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bf14564-2c89-4c30-abc1-06bddc04a0d6-utilities\") pod \"certified-operators-wxg6r\" (UID: \"3bf14564-2c89-4c30-abc1-06bddc04a0d6\") " pod="openshift-marketplace/certified-operators-wxg6r" Nov 24 07:39:39 crc kubenswrapper[4809]: I1124 07:39:39.199089 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bf14564-2c89-4c30-abc1-06bddc04a0d6-catalog-content\") pod \"certified-operators-wxg6r\" (UID: \"3bf14564-2c89-4c30-abc1-06bddc04a0d6\") " pod="openshift-marketplace/certified-operators-wxg6r" Nov 24 07:39:39 crc kubenswrapper[4809]: I1124 07:39:39.199413 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bf14564-2c89-4c30-abc1-06bddc04a0d6-utilities\") pod \"certified-operators-wxg6r\" (UID: \"3bf14564-2c89-4c30-abc1-06bddc04a0d6\") " pod="openshift-marketplace/certified-operators-wxg6r" Nov 24 07:39:39 crc kubenswrapper[4809]: I1124 07:39:39.199602 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-555xs\" (UniqueName: \"kubernetes.io/projected/3bf14564-2c89-4c30-abc1-06bddc04a0d6-kube-api-access-555xs\") pod \"certified-operators-wxg6r\" (UID: \"3bf14564-2c89-4c30-abc1-06bddc04a0d6\") " pod="openshift-marketplace/certified-operators-wxg6r" Nov 24 07:39:39 crc kubenswrapper[4809]: I1124 07:39:39.199625 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bf14564-2c89-4c30-abc1-06bddc04a0d6-catalog-content\") pod \"certified-operators-wxg6r\" (UID: \"3bf14564-2c89-4c30-abc1-06bddc04a0d6\") " pod="openshift-marketplace/certified-operators-wxg6r" Nov 24 07:39:39 crc kubenswrapper[4809]: I1124 07:39:39.199908 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bf14564-2c89-4c30-abc1-06bddc04a0d6-utilities\") pod \"certified-operators-wxg6r\" (UID: \"3bf14564-2c89-4c30-abc1-06bddc04a0d6\") " pod="openshift-marketplace/certified-operators-wxg6r" Nov 24 07:39:39 crc kubenswrapper[4809]: I1124 07:39:39.233810 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-555xs\" (UniqueName: \"kubernetes.io/projected/3bf14564-2c89-4c30-abc1-06bddc04a0d6-kube-api-access-555xs\") pod \"certified-operators-wxg6r\" (UID: \"3bf14564-2c89-4c30-abc1-06bddc04a0d6\") " pod="openshift-marketplace/certified-operators-wxg6r" Nov 24 07:39:39 crc kubenswrapper[4809]: I1124 07:39:39.320615 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wxg6r" Nov 24 07:39:39 crc kubenswrapper[4809]: I1124 07:39:39.408605 4809 generic.go:334] "Generic (PLEG): container finished" podID="17720a8a-d721-431b-8bc4-1efd5ac1a634" containerID="a817dfa8d48ec23e834dafca897b8994900272c83e4058bc3827c21765fb6dab" exitCode=0 Nov 24 07:39:39 crc kubenswrapper[4809]: I1124 07:39:39.408694 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nzgcn" event={"ID":"17720a8a-d721-431b-8bc4-1efd5ac1a634","Type":"ContainerDied","Data":"a817dfa8d48ec23e834dafca897b8994900272c83e4058bc3827c21765fb6dab"} Nov 24 07:39:39 crc kubenswrapper[4809]: I1124 07:39:39.408778 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nzgcn" event={"ID":"17720a8a-d721-431b-8bc4-1efd5ac1a634","Type":"ContainerStarted","Data":"731451a294191e19b88f501bb618937a39f3e22fcb584e52262a71458f3738f3"} Nov 24 07:39:39 crc kubenswrapper[4809]: I1124 07:39:39.931925 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wxg6r"] Nov 24 07:39:40 crc kubenswrapper[4809]: I1124 07:39:40.419991 4809 generic.go:334] "Generic (PLEG): container finished" podID="3bf14564-2c89-4c30-abc1-06bddc04a0d6" containerID="7fc41a24c5ffd37e4187750e8b1079f497bc778e5205f755c0b21663c312c9fe" exitCode=0 Nov 24 07:39:40 crc kubenswrapper[4809]: I1124 07:39:40.420059 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wxg6r" event={"ID":"3bf14564-2c89-4c30-abc1-06bddc04a0d6","Type":"ContainerDied","Data":"7fc41a24c5ffd37e4187750e8b1079f497bc778e5205f755c0b21663c312c9fe"} Nov 24 07:39:40 crc kubenswrapper[4809]: I1124 07:39:40.420156 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wxg6r" event={"ID":"3bf14564-2c89-4c30-abc1-06bddc04a0d6","Type":"ContainerStarted","Data":"c08ac90bf080fe1562b18fdd706dc0817773b9d61ba4c8950cd820da95814928"} Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.197000 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.198637 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.202154 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.202545 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.206016 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-dqvf8" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.206200 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.241538 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4f90573b-06dc-47d5-ab8b-5dab5503e96a-config-data\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.241640 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4f90573b-06dc-47d5-ab8b-5dab5503e96a-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.241700 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.241993 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4f90573b-06dc-47d5-ab8b-5dab5503e96a-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.242030 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4f90573b-06dc-47d5-ab8b-5dab5503e96a-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.242110 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4f90573b-06dc-47d5-ab8b-5dab5503e96a-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.242168 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f90573b-06dc-47d5-ab8b-5dab5503e96a-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.242194 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb6vh\" (UniqueName: \"kubernetes.io/projected/4f90573b-06dc-47d5-ab8b-5dab5503e96a-kube-api-access-tb6vh\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.242231 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4f90573b-06dc-47d5-ab8b-5dab5503e96a-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.258682 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.344342 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4f90573b-06dc-47d5-ab8b-5dab5503e96a-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.344795 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f90573b-06dc-47d5-ab8b-5dab5503e96a-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.344832 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb6vh\" (UniqueName: \"kubernetes.io/projected/4f90573b-06dc-47d5-ab8b-5dab5503e96a-kube-api-access-tb6vh\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.344878 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4f90573b-06dc-47d5-ab8b-5dab5503e96a-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.344938 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4f90573b-06dc-47d5-ab8b-5dab5503e96a-config-data\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.345117 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4f90573b-06dc-47d5-ab8b-5dab5503e96a-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.345254 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.345324 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4f90573b-06dc-47d5-ab8b-5dab5503e96a-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.345889 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4f90573b-06dc-47d5-ab8b-5dab5503e96a-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.346161 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.346242 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4f90573b-06dc-47d5-ab8b-5dab5503e96a-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.346583 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4f90573b-06dc-47d5-ab8b-5dab5503e96a-config-data\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.345355 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4f90573b-06dc-47d5-ab8b-5dab5503e96a-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.347016 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4f90573b-06dc-47d5-ab8b-5dab5503e96a-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.354029 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f90573b-06dc-47d5-ab8b-5dab5503e96a-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.354112 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4f90573b-06dc-47d5-ab8b-5dab5503e96a-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.358055 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4f90573b-06dc-47d5-ab8b-5dab5503e96a-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.362105 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb6vh\" (UniqueName: \"kubernetes.io/projected/4f90573b-06dc-47d5-ab8b-5dab5503e96a-kube-api-access-tb6vh\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.382930 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " pod="openstack/tempest-tests-tempest" Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.430422 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wxg6r" event={"ID":"3bf14564-2c89-4c30-abc1-06bddc04a0d6","Type":"ContainerStarted","Data":"7d4e0f226c9573ab407f7633dc6293469e24b8e2c74336d87558cbed4e956bbe"} Nov 24 07:39:41 crc kubenswrapper[4809]: I1124 07:39:41.592727 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 07:39:42 crc kubenswrapper[4809]: I1124 07:39:42.005387 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 24 07:39:42 crc kubenswrapper[4809]: W1124 07:39:42.026951 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f90573b_06dc_47d5_ab8b_5dab5503e96a.slice/crio-47000287f8bcad8ad972a79926af841302b45adb3f916bf0c837d7725cdc7551 WatchSource:0}: Error finding container 47000287f8bcad8ad972a79926af841302b45adb3f916bf0c837d7725cdc7551: Status 404 returned error can't find the container with id 47000287f8bcad8ad972a79926af841302b45adb3f916bf0c837d7725cdc7551 Nov 24 07:39:42 crc kubenswrapper[4809]: I1124 07:39:42.441394 4809 generic.go:334] "Generic (PLEG): container finished" podID="3bf14564-2c89-4c30-abc1-06bddc04a0d6" containerID="7d4e0f226c9573ab407f7633dc6293469e24b8e2c74336d87558cbed4e956bbe" exitCode=0 Nov 24 07:39:42 crc kubenswrapper[4809]: I1124 07:39:42.441824 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wxg6r" event={"ID":"3bf14564-2c89-4c30-abc1-06bddc04a0d6","Type":"ContainerDied","Data":"7d4e0f226c9573ab407f7633dc6293469e24b8e2c74336d87558cbed4e956bbe"} Nov 24 07:39:42 crc kubenswrapper[4809]: I1124 07:39:42.444440 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"4f90573b-06dc-47d5-ab8b-5dab5503e96a","Type":"ContainerStarted","Data":"47000287f8bcad8ad972a79926af841302b45adb3f916bf0c837d7725cdc7551"} Nov 24 07:39:43 crc kubenswrapper[4809]: I1124 07:39:43.464495 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wxg6r" event={"ID":"3bf14564-2c89-4c30-abc1-06bddc04a0d6","Type":"ContainerStarted","Data":"a860df032a2cc17415d86a11d71425f08bc6c7af34bd271c846742f413076612"} Nov 24 07:39:43 crc kubenswrapper[4809]: I1124 07:39:43.490570 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wxg6r" podStartSLOduration=3.045275358 podStartE2EDuration="5.490549768s" podCreationTimestamp="2025-11-24 07:39:38 +0000 UTC" firstStartedPulling="2025-11-24 07:39:40.423842956 +0000 UTC m=+2680.324434651" lastFinishedPulling="2025-11-24 07:39:42.869117356 +0000 UTC m=+2682.769709061" observedRunningTime="2025-11-24 07:39:43.485341589 +0000 UTC m=+2683.385933294" watchObservedRunningTime="2025-11-24 07:39:43.490549768 +0000 UTC m=+2683.391141473" Nov 24 07:39:48 crc kubenswrapper[4809]: I1124 07:39:48.043461 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:39:48 crc kubenswrapper[4809]: I1124 07:39:48.044056 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:39:48 crc kubenswrapper[4809]: I1124 07:39:48.044149 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 07:39:48 crc kubenswrapper[4809]: I1124 07:39:48.044928 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"630f55bb399f3bfd04e8309b3cf29adad95cffa6451d4ba31d28cc6a268df881"} pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:39:48 crc kubenswrapper[4809]: I1124 07:39:48.044995 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" containerID="cri-o://630f55bb399f3bfd04e8309b3cf29adad95cffa6451d4ba31d28cc6a268df881" gracePeriod=600 Nov 24 07:39:48 crc kubenswrapper[4809]: I1124 07:39:48.523590 4809 generic.go:334] "Generic (PLEG): container finished" podID="be8050d5-1aea-453e-a06e-87702b0e856b" containerID="630f55bb399f3bfd04e8309b3cf29adad95cffa6451d4ba31d28cc6a268df881" exitCode=0 Nov 24 07:39:48 crc kubenswrapper[4809]: I1124 07:39:48.523637 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerDied","Data":"630f55bb399f3bfd04e8309b3cf29adad95cffa6451d4ba31d28cc6a268df881"} Nov 24 07:39:48 crc kubenswrapper[4809]: I1124 07:39:48.523935 4809 scope.go:117] "RemoveContainer" containerID="b910815982bda57f6203972a26ac7c69c2f05c4f95705cbaab87c197220f1add" Nov 24 07:39:49 crc kubenswrapper[4809]: I1124 07:39:49.323811 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wxg6r" Nov 24 07:39:49 crc kubenswrapper[4809]: I1124 07:39:49.323866 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wxg6r" Nov 24 07:39:49 crc kubenswrapper[4809]: I1124 07:39:49.391825 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wxg6r" Nov 24 07:39:49 crc kubenswrapper[4809]: I1124 07:39:49.595644 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wxg6r" Nov 24 07:39:49 crc kubenswrapper[4809]: I1124 07:39:49.643648 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wxg6r"] Nov 24 07:39:51 crc kubenswrapper[4809]: I1124 07:39:51.560359 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wxg6r" podUID="3bf14564-2c89-4c30-abc1-06bddc04a0d6" containerName="registry-server" containerID="cri-o://a860df032a2cc17415d86a11d71425f08bc6c7af34bd271c846742f413076612" gracePeriod=2 Nov 24 07:39:52 crc kubenswrapper[4809]: I1124 07:39:52.573268 4809 generic.go:334] "Generic (PLEG): container finished" podID="3bf14564-2c89-4c30-abc1-06bddc04a0d6" containerID="a860df032a2cc17415d86a11d71425f08bc6c7af34bd271c846742f413076612" exitCode=0 Nov 24 07:39:52 crc kubenswrapper[4809]: I1124 07:39:52.573764 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wxg6r" event={"ID":"3bf14564-2c89-4c30-abc1-06bddc04a0d6","Type":"ContainerDied","Data":"a860df032a2cc17415d86a11d71425f08bc6c7af34bd271c846742f413076612"} Nov 24 07:39:59 crc kubenswrapper[4809]: E1124 07:39:59.321720 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a860df032a2cc17415d86a11d71425f08bc6c7af34bd271c846742f413076612 is running failed: container process not found" containerID="a860df032a2cc17415d86a11d71425f08bc6c7af34bd271c846742f413076612" cmd=["grpc_health_probe","-addr=:50051"] Nov 24 07:39:59 crc kubenswrapper[4809]: E1124 07:39:59.322681 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a860df032a2cc17415d86a11d71425f08bc6c7af34bd271c846742f413076612 is running failed: container process not found" containerID="a860df032a2cc17415d86a11d71425f08bc6c7af34bd271c846742f413076612" cmd=["grpc_health_probe","-addr=:50051"] Nov 24 07:39:59 crc kubenswrapper[4809]: E1124 07:39:59.322980 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a860df032a2cc17415d86a11d71425f08bc6c7af34bd271c846742f413076612 is running failed: container process not found" containerID="a860df032a2cc17415d86a11d71425f08bc6c7af34bd271c846742f413076612" cmd=["grpc_health_probe","-addr=:50051"] Nov 24 07:39:59 crc kubenswrapper[4809]: E1124 07:39:59.323026 4809 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a860df032a2cc17415d86a11d71425f08bc6c7af34bd271c846742f413076612 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-wxg6r" podUID="3bf14564-2c89-4c30-abc1-06bddc04a0d6" containerName="registry-server" Nov 24 07:40:09 crc kubenswrapper[4809]: E1124 07:40:09.324516 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a860df032a2cc17415d86a11d71425f08bc6c7af34bd271c846742f413076612 is running failed: container process not found" containerID="a860df032a2cc17415d86a11d71425f08bc6c7af34bd271c846742f413076612" cmd=["grpc_health_probe","-addr=:50051"] Nov 24 07:40:09 crc kubenswrapper[4809]: E1124 07:40:09.325397 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a860df032a2cc17415d86a11d71425f08bc6c7af34bd271c846742f413076612 is running failed: container process not found" containerID="a860df032a2cc17415d86a11d71425f08bc6c7af34bd271c846742f413076612" cmd=["grpc_health_probe","-addr=:50051"] Nov 24 07:40:09 crc kubenswrapper[4809]: E1124 07:40:09.325949 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a860df032a2cc17415d86a11d71425f08bc6c7af34bd271c846742f413076612 is running failed: container process not found" containerID="a860df032a2cc17415d86a11d71425f08bc6c7af34bd271c846742f413076612" cmd=["grpc_health_probe","-addr=:50051"] Nov 24 07:40:09 crc kubenswrapper[4809]: E1124 07:40:09.326006 4809 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a860df032a2cc17415d86a11d71425f08bc6c7af34bd271c846742f413076612 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-wxg6r" podUID="3bf14564-2c89-4c30-abc1-06bddc04a0d6" containerName="registry-server" Nov 24 07:40:11 crc kubenswrapper[4809]: E1124 07:40:11.736014 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Nov 24 07:40:11 crc kubenswrapper[4809]: E1124 07:40:11.736508 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tb6vh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(4f90573b-06dc-47d5-ab8b-5dab5503e96a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 07:40:11 crc kubenswrapper[4809]: E1124 07:40:11.737701 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="4f90573b-06dc-47d5-ab8b-5dab5503e96a" Nov 24 07:40:11 crc kubenswrapper[4809]: E1124 07:40:11.800633 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="4f90573b-06dc-47d5-ab8b-5dab5503e96a" Nov 24 07:40:12 crc kubenswrapper[4809]: I1124 07:40:12.023333 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wxg6r" Nov 24 07:40:12 crc kubenswrapper[4809]: I1124 07:40:12.148138 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bf14564-2c89-4c30-abc1-06bddc04a0d6-utilities\") pod \"3bf14564-2c89-4c30-abc1-06bddc04a0d6\" (UID: \"3bf14564-2c89-4c30-abc1-06bddc04a0d6\") " Nov 24 07:40:12 crc kubenswrapper[4809]: I1124 07:40:12.148330 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-555xs\" (UniqueName: \"kubernetes.io/projected/3bf14564-2c89-4c30-abc1-06bddc04a0d6-kube-api-access-555xs\") pod \"3bf14564-2c89-4c30-abc1-06bddc04a0d6\" (UID: \"3bf14564-2c89-4c30-abc1-06bddc04a0d6\") " Nov 24 07:40:12 crc kubenswrapper[4809]: I1124 07:40:12.148381 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bf14564-2c89-4c30-abc1-06bddc04a0d6-catalog-content\") pod \"3bf14564-2c89-4c30-abc1-06bddc04a0d6\" (UID: \"3bf14564-2c89-4c30-abc1-06bddc04a0d6\") " Nov 24 07:40:12 crc kubenswrapper[4809]: I1124 07:40:12.148742 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bf14564-2c89-4c30-abc1-06bddc04a0d6-utilities" (OuterVolumeSpecName: "utilities") pod "3bf14564-2c89-4c30-abc1-06bddc04a0d6" (UID: "3bf14564-2c89-4c30-abc1-06bddc04a0d6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:40:12 crc kubenswrapper[4809]: I1124 07:40:12.154912 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bf14564-2c89-4c30-abc1-06bddc04a0d6-kube-api-access-555xs" (OuterVolumeSpecName: "kube-api-access-555xs") pod "3bf14564-2c89-4c30-abc1-06bddc04a0d6" (UID: "3bf14564-2c89-4c30-abc1-06bddc04a0d6"). InnerVolumeSpecName "kube-api-access-555xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:40:12 crc kubenswrapper[4809]: I1124 07:40:12.202918 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bf14564-2c89-4c30-abc1-06bddc04a0d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3bf14564-2c89-4c30-abc1-06bddc04a0d6" (UID: "3bf14564-2c89-4c30-abc1-06bddc04a0d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:40:12 crc kubenswrapper[4809]: I1124 07:40:12.250207 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-555xs\" (UniqueName: \"kubernetes.io/projected/3bf14564-2c89-4c30-abc1-06bddc04a0d6-kube-api-access-555xs\") on node \"crc\" DevicePath \"\"" Nov 24 07:40:12 crc kubenswrapper[4809]: I1124 07:40:12.250238 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bf14564-2c89-4c30-abc1-06bddc04a0d6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:40:12 crc kubenswrapper[4809]: I1124 07:40:12.250250 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bf14564-2c89-4c30-abc1-06bddc04a0d6-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:40:12 crc kubenswrapper[4809]: I1124 07:40:12.813337 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerStarted","Data":"1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759"} Nov 24 07:40:12 crc kubenswrapper[4809]: I1124 07:40:12.816191 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wxg6r" Nov 24 07:40:12 crc kubenswrapper[4809]: I1124 07:40:12.816183 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wxg6r" event={"ID":"3bf14564-2c89-4c30-abc1-06bddc04a0d6","Type":"ContainerDied","Data":"c08ac90bf080fe1562b18fdd706dc0817773b9d61ba4c8950cd820da95814928"} Nov 24 07:40:12 crc kubenswrapper[4809]: I1124 07:40:12.816312 4809 scope.go:117] "RemoveContainer" containerID="a860df032a2cc17415d86a11d71425f08bc6c7af34bd271c846742f413076612" Nov 24 07:40:12 crc kubenswrapper[4809]: I1124 07:40:12.818140 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nzgcn" event={"ID":"17720a8a-d721-431b-8bc4-1efd5ac1a634","Type":"ContainerStarted","Data":"0b5077000ba246a32d68268da364aff778d534de646f51f7f2141f9beae210fa"} Nov 24 07:40:12 crc kubenswrapper[4809]: I1124 07:40:12.857394 4809 scope.go:117] "RemoveContainer" containerID="7d4e0f226c9573ab407f7633dc6293469e24b8e2c74336d87558cbed4e956bbe" Nov 24 07:40:12 crc kubenswrapper[4809]: I1124 07:40:12.877727 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wxg6r"] Nov 24 07:40:12 crc kubenswrapper[4809]: I1124 07:40:12.886452 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wxg6r"] Nov 24 07:40:12 crc kubenswrapper[4809]: I1124 07:40:12.903885 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bf14564-2c89-4c30-abc1-06bddc04a0d6" path="/var/lib/kubelet/pods/3bf14564-2c89-4c30-abc1-06bddc04a0d6/volumes" Nov 24 07:40:12 crc kubenswrapper[4809]: I1124 07:40:12.908626 4809 scope.go:117] "RemoveContainer" containerID="7fc41a24c5ffd37e4187750e8b1079f497bc778e5205f755c0b21663c312c9fe" Nov 24 07:40:19 crc kubenswrapper[4809]: I1124 07:40:19.878950 4809 generic.go:334] "Generic (PLEG): container finished" podID="17720a8a-d721-431b-8bc4-1efd5ac1a634" containerID="0b5077000ba246a32d68268da364aff778d534de646f51f7f2141f9beae210fa" exitCode=0 Nov 24 07:40:19 crc kubenswrapper[4809]: I1124 07:40:19.879052 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nzgcn" event={"ID":"17720a8a-d721-431b-8bc4-1efd5ac1a634","Type":"ContainerDied","Data":"0b5077000ba246a32d68268da364aff778d534de646f51f7f2141f9beae210fa"} Nov 24 07:40:20 crc kubenswrapper[4809]: I1124 07:40:20.910240 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nzgcn" event={"ID":"17720a8a-d721-431b-8bc4-1efd5ac1a634","Type":"ContainerStarted","Data":"ffa88530e4629f3b2004da049c856d0df53ee2705c6e826b0deddee588b84775"} Nov 24 07:40:20 crc kubenswrapper[4809]: I1124 07:40:20.938040 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nzgcn" podStartSLOduration=3.053118895 podStartE2EDuration="43.938019286s" podCreationTimestamp="2025-11-24 07:39:37 +0000 UTC" firstStartedPulling="2025-11-24 07:39:39.414988078 +0000 UTC m=+2679.315579783" lastFinishedPulling="2025-11-24 07:40:20.299888469 +0000 UTC m=+2720.200480174" observedRunningTime="2025-11-24 07:40:20.926417266 +0000 UTC m=+2720.827008971" watchObservedRunningTime="2025-11-24 07:40:20.938019286 +0000 UTC m=+2720.838610991" Nov 24 07:40:23 crc kubenswrapper[4809]: I1124 07:40:23.290945 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 24 07:40:24 crc kubenswrapper[4809]: I1124 07:40:24.924844 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"4f90573b-06dc-47d5-ab8b-5dab5503e96a","Type":"ContainerStarted","Data":"b3eedb3a6223893f8c6be0a1b50d68784be92462929ec43f020a7f4833601c88"} Nov 24 07:40:24 crc kubenswrapper[4809]: I1124 07:40:24.950685 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.695699866 podStartE2EDuration="44.950666818s" podCreationTimestamp="2025-11-24 07:39:40 +0000 UTC" firstStartedPulling="2025-11-24 07:39:42.028920024 +0000 UTC m=+2681.929511729" lastFinishedPulling="2025-11-24 07:40:23.283886976 +0000 UTC m=+2723.184478681" observedRunningTime="2025-11-24 07:40:24.944608757 +0000 UTC m=+2724.845200472" watchObservedRunningTime="2025-11-24 07:40:24.950666818 +0000 UTC m=+2724.851258533" Nov 24 07:40:27 crc kubenswrapper[4809]: I1124 07:40:27.987605 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nzgcn" Nov 24 07:40:27 crc kubenswrapper[4809]: I1124 07:40:27.988136 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nzgcn" Nov 24 07:40:29 crc kubenswrapper[4809]: I1124 07:40:29.041237 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nzgcn" podUID="17720a8a-d721-431b-8bc4-1efd5ac1a634" containerName="registry-server" probeResult="failure" output=< Nov 24 07:40:29 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Nov 24 07:40:29 crc kubenswrapper[4809]: > Nov 24 07:40:39 crc kubenswrapper[4809]: I1124 07:40:39.036826 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nzgcn" podUID="17720a8a-d721-431b-8bc4-1efd5ac1a634" containerName="registry-server" probeResult="failure" output=< Nov 24 07:40:39 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Nov 24 07:40:39 crc kubenswrapper[4809]: > Nov 24 07:40:48 crc kubenswrapper[4809]: I1124 07:40:48.034924 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nzgcn" Nov 24 07:40:48 crc kubenswrapper[4809]: I1124 07:40:48.086668 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nzgcn" Nov 24 07:40:48 crc kubenswrapper[4809]: I1124 07:40:48.161382 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nzgcn"] Nov 24 07:40:48 crc kubenswrapper[4809]: I1124 07:40:48.273109 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w5ghj"] Nov 24 07:40:48 crc kubenswrapper[4809]: I1124 07:40:48.273665 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-w5ghj" podUID="2e09c138-944f-42fa-b52f-cc3cf1395226" containerName="registry-server" containerID="cri-o://0737d3bdf4f5ef9427ec877a4078201a87fc15cc6ea2b40027b831bea5805efe" gracePeriod=2 Nov 24 07:40:48 crc kubenswrapper[4809]: I1124 07:40:48.708299 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w5ghj" Nov 24 07:40:48 crc kubenswrapper[4809]: I1124 07:40:48.891151 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqcq8\" (UniqueName: \"kubernetes.io/projected/2e09c138-944f-42fa-b52f-cc3cf1395226-kube-api-access-tqcq8\") pod \"2e09c138-944f-42fa-b52f-cc3cf1395226\" (UID: \"2e09c138-944f-42fa-b52f-cc3cf1395226\") " Nov 24 07:40:48 crc kubenswrapper[4809]: I1124 07:40:48.891622 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e09c138-944f-42fa-b52f-cc3cf1395226-catalog-content\") pod \"2e09c138-944f-42fa-b52f-cc3cf1395226\" (UID: \"2e09c138-944f-42fa-b52f-cc3cf1395226\") " Nov 24 07:40:48 crc kubenswrapper[4809]: I1124 07:40:48.891691 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e09c138-944f-42fa-b52f-cc3cf1395226-utilities\") pod \"2e09c138-944f-42fa-b52f-cc3cf1395226\" (UID: \"2e09c138-944f-42fa-b52f-cc3cf1395226\") " Nov 24 07:40:48 crc kubenswrapper[4809]: I1124 07:40:48.893155 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e09c138-944f-42fa-b52f-cc3cf1395226-utilities" (OuterVolumeSpecName: "utilities") pod "2e09c138-944f-42fa-b52f-cc3cf1395226" (UID: "2e09c138-944f-42fa-b52f-cc3cf1395226"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:40:48 crc kubenswrapper[4809]: I1124 07:40:48.901775 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e09c138-944f-42fa-b52f-cc3cf1395226-kube-api-access-tqcq8" (OuterVolumeSpecName: "kube-api-access-tqcq8") pod "2e09c138-944f-42fa-b52f-cc3cf1395226" (UID: "2e09c138-944f-42fa-b52f-cc3cf1395226"). InnerVolumeSpecName "kube-api-access-tqcq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:40:48 crc kubenswrapper[4809]: I1124 07:40:48.995036 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e09c138-944f-42fa-b52f-cc3cf1395226-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:40:48 crc kubenswrapper[4809]: I1124 07:40:48.995072 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqcq8\" (UniqueName: \"kubernetes.io/projected/2e09c138-944f-42fa-b52f-cc3cf1395226-kube-api-access-tqcq8\") on node \"crc\" DevicePath \"\"" Nov 24 07:40:48 crc kubenswrapper[4809]: I1124 07:40:48.995583 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e09c138-944f-42fa-b52f-cc3cf1395226-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2e09c138-944f-42fa-b52f-cc3cf1395226" (UID: "2e09c138-944f-42fa-b52f-cc3cf1395226"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:40:49 crc kubenswrapper[4809]: I1124 07:40:49.096312 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e09c138-944f-42fa-b52f-cc3cf1395226-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:40:49 crc kubenswrapper[4809]: I1124 07:40:49.133251 4809 generic.go:334] "Generic (PLEG): container finished" podID="2e09c138-944f-42fa-b52f-cc3cf1395226" containerID="0737d3bdf4f5ef9427ec877a4078201a87fc15cc6ea2b40027b831bea5805efe" exitCode=0 Nov 24 07:40:49 crc kubenswrapper[4809]: I1124 07:40:49.133302 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w5ghj" Nov 24 07:40:49 crc kubenswrapper[4809]: I1124 07:40:49.133369 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5ghj" event={"ID":"2e09c138-944f-42fa-b52f-cc3cf1395226","Type":"ContainerDied","Data":"0737d3bdf4f5ef9427ec877a4078201a87fc15cc6ea2b40027b831bea5805efe"} Nov 24 07:40:49 crc kubenswrapper[4809]: I1124 07:40:49.133448 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5ghj" event={"ID":"2e09c138-944f-42fa-b52f-cc3cf1395226","Type":"ContainerDied","Data":"d07072d5b359b48c377cf55aa2ce6d55316d1fd5403ee37c7884d3f5817b7690"} Nov 24 07:40:49 crc kubenswrapper[4809]: I1124 07:40:49.133482 4809 scope.go:117] "RemoveContainer" containerID="0737d3bdf4f5ef9427ec877a4078201a87fc15cc6ea2b40027b831bea5805efe" Nov 24 07:40:49 crc kubenswrapper[4809]: I1124 07:40:49.165378 4809 scope.go:117] "RemoveContainer" containerID="0264b58e2db49abf24a5d4368e0c97913c433a9e81226f9b7e1b742d988d08cd" Nov 24 07:40:49 crc kubenswrapper[4809]: I1124 07:40:49.170018 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w5ghj"] Nov 24 07:40:49 crc kubenswrapper[4809]: I1124 07:40:49.181392 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-w5ghj"] Nov 24 07:40:49 crc kubenswrapper[4809]: I1124 07:40:49.182507 4809 scope.go:117] "RemoveContainer" containerID="245926ab3d8a7793dddb426aaa948f234738df85b6238085ecdda67451f0e0bf" Nov 24 07:40:49 crc kubenswrapper[4809]: I1124 07:40:49.227006 4809 scope.go:117] "RemoveContainer" containerID="0737d3bdf4f5ef9427ec877a4078201a87fc15cc6ea2b40027b831bea5805efe" Nov 24 07:40:49 crc kubenswrapper[4809]: E1124 07:40:49.227500 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0737d3bdf4f5ef9427ec877a4078201a87fc15cc6ea2b40027b831bea5805efe\": container with ID starting with 0737d3bdf4f5ef9427ec877a4078201a87fc15cc6ea2b40027b831bea5805efe not found: ID does not exist" containerID="0737d3bdf4f5ef9427ec877a4078201a87fc15cc6ea2b40027b831bea5805efe" Nov 24 07:40:49 crc kubenswrapper[4809]: I1124 07:40:49.227544 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0737d3bdf4f5ef9427ec877a4078201a87fc15cc6ea2b40027b831bea5805efe"} err="failed to get container status \"0737d3bdf4f5ef9427ec877a4078201a87fc15cc6ea2b40027b831bea5805efe\": rpc error: code = NotFound desc = could not find container \"0737d3bdf4f5ef9427ec877a4078201a87fc15cc6ea2b40027b831bea5805efe\": container with ID starting with 0737d3bdf4f5ef9427ec877a4078201a87fc15cc6ea2b40027b831bea5805efe not found: ID does not exist" Nov 24 07:40:49 crc kubenswrapper[4809]: I1124 07:40:49.227577 4809 scope.go:117] "RemoveContainer" containerID="0264b58e2db49abf24a5d4368e0c97913c433a9e81226f9b7e1b742d988d08cd" Nov 24 07:40:49 crc kubenswrapper[4809]: E1124 07:40:49.228317 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0264b58e2db49abf24a5d4368e0c97913c433a9e81226f9b7e1b742d988d08cd\": container with ID starting with 0264b58e2db49abf24a5d4368e0c97913c433a9e81226f9b7e1b742d988d08cd not found: ID does not exist" containerID="0264b58e2db49abf24a5d4368e0c97913c433a9e81226f9b7e1b742d988d08cd" Nov 24 07:40:49 crc kubenswrapper[4809]: I1124 07:40:49.228373 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0264b58e2db49abf24a5d4368e0c97913c433a9e81226f9b7e1b742d988d08cd"} err="failed to get container status \"0264b58e2db49abf24a5d4368e0c97913c433a9e81226f9b7e1b742d988d08cd\": rpc error: code = NotFound desc = could not find container \"0264b58e2db49abf24a5d4368e0c97913c433a9e81226f9b7e1b742d988d08cd\": container with ID starting with 0264b58e2db49abf24a5d4368e0c97913c433a9e81226f9b7e1b742d988d08cd not found: ID does not exist" Nov 24 07:40:49 crc kubenswrapper[4809]: I1124 07:40:49.228414 4809 scope.go:117] "RemoveContainer" containerID="245926ab3d8a7793dddb426aaa948f234738df85b6238085ecdda67451f0e0bf" Nov 24 07:40:49 crc kubenswrapper[4809]: E1124 07:40:49.228722 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"245926ab3d8a7793dddb426aaa948f234738df85b6238085ecdda67451f0e0bf\": container with ID starting with 245926ab3d8a7793dddb426aaa948f234738df85b6238085ecdda67451f0e0bf not found: ID does not exist" containerID="245926ab3d8a7793dddb426aaa948f234738df85b6238085ecdda67451f0e0bf" Nov 24 07:40:49 crc kubenswrapper[4809]: I1124 07:40:49.228746 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"245926ab3d8a7793dddb426aaa948f234738df85b6238085ecdda67451f0e0bf"} err="failed to get container status \"245926ab3d8a7793dddb426aaa948f234738df85b6238085ecdda67451f0e0bf\": rpc error: code = NotFound desc = could not find container \"245926ab3d8a7793dddb426aaa948f234738df85b6238085ecdda67451f0e0bf\": container with ID starting with 245926ab3d8a7793dddb426aaa948f234738df85b6238085ecdda67451f0e0bf not found: ID does not exist" Nov 24 07:40:50 crc kubenswrapper[4809]: I1124 07:40:50.902251 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e09c138-944f-42fa-b52f-cc3cf1395226" path="/var/lib/kubelet/pods/2e09c138-944f-42fa-b52f-cc3cf1395226/volumes" Nov 24 07:42:18 crc kubenswrapper[4809]: I1124 07:42:18.042811 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:42:18 crc kubenswrapper[4809]: I1124 07:42:18.043262 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:42:27 crc kubenswrapper[4809]: I1124 07:42:27.402541 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ff8j2"] Nov 24 07:42:27 crc kubenswrapper[4809]: E1124 07:42:27.403444 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e09c138-944f-42fa-b52f-cc3cf1395226" containerName="extract-utilities" Nov 24 07:42:27 crc kubenswrapper[4809]: I1124 07:42:27.403457 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e09c138-944f-42fa-b52f-cc3cf1395226" containerName="extract-utilities" Nov 24 07:42:27 crc kubenswrapper[4809]: E1124 07:42:27.403466 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e09c138-944f-42fa-b52f-cc3cf1395226" containerName="extract-content" Nov 24 07:42:27 crc kubenswrapper[4809]: I1124 07:42:27.403471 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e09c138-944f-42fa-b52f-cc3cf1395226" containerName="extract-content" Nov 24 07:42:27 crc kubenswrapper[4809]: E1124 07:42:27.403483 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bf14564-2c89-4c30-abc1-06bddc04a0d6" containerName="extract-content" Nov 24 07:42:27 crc kubenswrapper[4809]: I1124 07:42:27.403489 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bf14564-2c89-4c30-abc1-06bddc04a0d6" containerName="extract-content" Nov 24 07:42:27 crc kubenswrapper[4809]: E1124 07:42:27.403507 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bf14564-2c89-4c30-abc1-06bddc04a0d6" containerName="registry-server" Nov 24 07:42:27 crc kubenswrapper[4809]: I1124 07:42:27.403513 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bf14564-2c89-4c30-abc1-06bddc04a0d6" containerName="registry-server" Nov 24 07:42:27 crc kubenswrapper[4809]: E1124 07:42:27.403522 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bf14564-2c89-4c30-abc1-06bddc04a0d6" containerName="extract-utilities" Nov 24 07:42:27 crc kubenswrapper[4809]: I1124 07:42:27.403527 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bf14564-2c89-4c30-abc1-06bddc04a0d6" containerName="extract-utilities" Nov 24 07:42:27 crc kubenswrapper[4809]: E1124 07:42:27.403542 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e09c138-944f-42fa-b52f-cc3cf1395226" containerName="registry-server" Nov 24 07:42:27 crc kubenswrapper[4809]: I1124 07:42:27.403547 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e09c138-944f-42fa-b52f-cc3cf1395226" containerName="registry-server" Nov 24 07:42:27 crc kubenswrapper[4809]: I1124 07:42:27.403713 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e09c138-944f-42fa-b52f-cc3cf1395226" containerName="registry-server" Nov 24 07:42:27 crc kubenswrapper[4809]: I1124 07:42:27.403739 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bf14564-2c89-4c30-abc1-06bddc04a0d6" containerName="registry-server" Nov 24 07:42:27 crc kubenswrapper[4809]: I1124 07:42:27.404946 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ff8j2" Nov 24 07:42:27 crc kubenswrapper[4809]: I1124 07:42:27.424261 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ff8j2"] Nov 24 07:42:27 crc kubenswrapper[4809]: I1124 07:42:27.513181 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzsjb\" (UniqueName: \"kubernetes.io/projected/2408c6a0-ae42-4da0-a8d8-262db680df45-kube-api-access-xzsjb\") pod \"redhat-marketplace-ff8j2\" (UID: \"2408c6a0-ae42-4da0-a8d8-262db680df45\") " pod="openshift-marketplace/redhat-marketplace-ff8j2" Nov 24 07:42:27 crc kubenswrapper[4809]: I1124 07:42:27.513599 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2408c6a0-ae42-4da0-a8d8-262db680df45-utilities\") pod \"redhat-marketplace-ff8j2\" (UID: \"2408c6a0-ae42-4da0-a8d8-262db680df45\") " pod="openshift-marketplace/redhat-marketplace-ff8j2" Nov 24 07:42:27 crc kubenswrapper[4809]: I1124 07:42:27.513976 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2408c6a0-ae42-4da0-a8d8-262db680df45-catalog-content\") pod \"redhat-marketplace-ff8j2\" (UID: \"2408c6a0-ae42-4da0-a8d8-262db680df45\") " pod="openshift-marketplace/redhat-marketplace-ff8j2" Nov 24 07:42:27 crc kubenswrapper[4809]: I1124 07:42:27.615696 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2408c6a0-ae42-4da0-a8d8-262db680df45-utilities\") pod \"redhat-marketplace-ff8j2\" (UID: \"2408c6a0-ae42-4da0-a8d8-262db680df45\") " pod="openshift-marketplace/redhat-marketplace-ff8j2" Nov 24 07:42:27 crc kubenswrapper[4809]: I1124 07:42:27.615823 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2408c6a0-ae42-4da0-a8d8-262db680df45-catalog-content\") pod \"redhat-marketplace-ff8j2\" (UID: \"2408c6a0-ae42-4da0-a8d8-262db680df45\") " pod="openshift-marketplace/redhat-marketplace-ff8j2" Nov 24 07:42:27 crc kubenswrapper[4809]: I1124 07:42:27.615876 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzsjb\" (UniqueName: \"kubernetes.io/projected/2408c6a0-ae42-4da0-a8d8-262db680df45-kube-api-access-xzsjb\") pod \"redhat-marketplace-ff8j2\" (UID: \"2408c6a0-ae42-4da0-a8d8-262db680df45\") " pod="openshift-marketplace/redhat-marketplace-ff8j2" Nov 24 07:42:27 crc kubenswrapper[4809]: I1124 07:42:27.616405 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2408c6a0-ae42-4da0-a8d8-262db680df45-catalog-content\") pod \"redhat-marketplace-ff8j2\" (UID: \"2408c6a0-ae42-4da0-a8d8-262db680df45\") " pod="openshift-marketplace/redhat-marketplace-ff8j2" Nov 24 07:42:27 crc kubenswrapper[4809]: I1124 07:42:27.616481 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2408c6a0-ae42-4da0-a8d8-262db680df45-utilities\") pod \"redhat-marketplace-ff8j2\" (UID: \"2408c6a0-ae42-4da0-a8d8-262db680df45\") " pod="openshift-marketplace/redhat-marketplace-ff8j2" Nov 24 07:42:27 crc kubenswrapper[4809]: I1124 07:42:27.645745 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzsjb\" (UniqueName: \"kubernetes.io/projected/2408c6a0-ae42-4da0-a8d8-262db680df45-kube-api-access-xzsjb\") pod \"redhat-marketplace-ff8j2\" (UID: \"2408c6a0-ae42-4da0-a8d8-262db680df45\") " pod="openshift-marketplace/redhat-marketplace-ff8j2" Nov 24 07:42:27 crc kubenswrapper[4809]: I1124 07:42:27.727145 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ff8j2" Nov 24 07:42:28 crc kubenswrapper[4809]: I1124 07:42:28.182358 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ff8j2"] Nov 24 07:42:29 crc kubenswrapper[4809]: I1124 07:42:29.094684 4809 generic.go:334] "Generic (PLEG): container finished" podID="2408c6a0-ae42-4da0-a8d8-262db680df45" containerID="ac2aef5b94e892d15507bd4855919bb279fd5dcdbfd4f519563ff7c49faf4cea" exitCode=0 Nov 24 07:42:29 crc kubenswrapper[4809]: I1124 07:42:29.094765 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ff8j2" event={"ID":"2408c6a0-ae42-4da0-a8d8-262db680df45","Type":"ContainerDied","Data":"ac2aef5b94e892d15507bd4855919bb279fd5dcdbfd4f519563ff7c49faf4cea"} Nov 24 07:42:29 crc kubenswrapper[4809]: I1124 07:42:29.094986 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ff8j2" event={"ID":"2408c6a0-ae42-4da0-a8d8-262db680df45","Type":"ContainerStarted","Data":"b7541a7b4d560940e36387a4af2f85404c194e647b15a386706a39d16411a535"} Nov 24 07:42:30 crc kubenswrapper[4809]: I1124 07:42:30.109120 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ff8j2" event={"ID":"2408c6a0-ae42-4da0-a8d8-262db680df45","Type":"ContainerStarted","Data":"fb6839ee6d8f8b5f210bc7968ac75e9c84d5ed0566a51c5fee08e3ed173dd9e7"} Nov 24 07:42:31 crc kubenswrapper[4809]: I1124 07:42:31.125052 4809 generic.go:334] "Generic (PLEG): container finished" podID="2408c6a0-ae42-4da0-a8d8-262db680df45" containerID="fb6839ee6d8f8b5f210bc7968ac75e9c84d5ed0566a51c5fee08e3ed173dd9e7" exitCode=0 Nov 24 07:42:31 crc kubenswrapper[4809]: I1124 07:42:31.125118 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ff8j2" event={"ID":"2408c6a0-ae42-4da0-a8d8-262db680df45","Type":"ContainerDied","Data":"fb6839ee6d8f8b5f210bc7968ac75e9c84d5ed0566a51c5fee08e3ed173dd9e7"} Nov 24 07:42:32 crc kubenswrapper[4809]: I1124 07:42:32.135249 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ff8j2" event={"ID":"2408c6a0-ae42-4da0-a8d8-262db680df45","Type":"ContainerStarted","Data":"e126d8f8270a0862f6413c7b391e01eec779a701ff16f0b15f10f459b963ef06"} Nov 24 07:42:32 crc kubenswrapper[4809]: I1124 07:42:32.156423 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ff8j2" podStartSLOduration=2.4696066 podStartE2EDuration="5.156404039s" podCreationTimestamp="2025-11-24 07:42:27 +0000 UTC" firstStartedPulling="2025-11-24 07:42:29.096377413 +0000 UTC m=+2848.996969118" lastFinishedPulling="2025-11-24 07:42:31.783174852 +0000 UTC m=+2851.683766557" observedRunningTime="2025-11-24 07:42:32.154166945 +0000 UTC m=+2852.054758650" watchObservedRunningTime="2025-11-24 07:42:32.156404039 +0000 UTC m=+2852.056995744" Nov 24 07:42:37 crc kubenswrapper[4809]: I1124 07:42:37.728460 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ff8j2" Nov 24 07:42:37 crc kubenswrapper[4809]: I1124 07:42:37.729157 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ff8j2" Nov 24 07:42:37 crc kubenswrapper[4809]: I1124 07:42:37.782667 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ff8j2" Nov 24 07:42:38 crc kubenswrapper[4809]: I1124 07:42:38.261521 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ff8j2" Nov 24 07:42:40 crc kubenswrapper[4809]: I1124 07:42:40.601873 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ff8j2"] Nov 24 07:42:40 crc kubenswrapper[4809]: I1124 07:42:40.604020 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ff8j2" podUID="2408c6a0-ae42-4da0-a8d8-262db680df45" containerName="registry-server" containerID="cri-o://e126d8f8270a0862f6413c7b391e01eec779a701ff16f0b15f10f459b963ef06" gracePeriod=2 Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.123797 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ff8j2" Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.168602 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2408c6a0-ae42-4da0-a8d8-262db680df45-catalog-content\") pod \"2408c6a0-ae42-4da0-a8d8-262db680df45\" (UID: \"2408c6a0-ae42-4da0-a8d8-262db680df45\") " Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.168742 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzsjb\" (UniqueName: \"kubernetes.io/projected/2408c6a0-ae42-4da0-a8d8-262db680df45-kube-api-access-xzsjb\") pod \"2408c6a0-ae42-4da0-a8d8-262db680df45\" (UID: \"2408c6a0-ae42-4da0-a8d8-262db680df45\") " Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.168784 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2408c6a0-ae42-4da0-a8d8-262db680df45-utilities\") pod \"2408c6a0-ae42-4da0-a8d8-262db680df45\" (UID: \"2408c6a0-ae42-4da0-a8d8-262db680df45\") " Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.169556 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2408c6a0-ae42-4da0-a8d8-262db680df45-utilities" (OuterVolumeSpecName: "utilities") pod "2408c6a0-ae42-4da0-a8d8-262db680df45" (UID: "2408c6a0-ae42-4da0-a8d8-262db680df45"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.180469 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2408c6a0-ae42-4da0-a8d8-262db680df45-kube-api-access-xzsjb" (OuterVolumeSpecName: "kube-api-access-xzsjb") pod "2408c6a0-ae42-4da0-a8d8-262db680df45" (UID: "2408c6a0-ae42-4da0-a8d8-262db680df45"). InnerVolumeSpecName "kube-api-access-xzsjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.185703 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2408c6a0-ae42-4da0-a8d8-262db680df45-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2408c6a0-ae42-4da0-a8d8-262db680df45" (UID: "2408c6a0-ae42-4da0-a8d8-262db680df45"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.223907 4809 generic.go:334] "Generic (PLEG): container finished" podID="2408c6a0-ae42-4da0-a8d8-262db680df45" containerID="e126d8f8270a0862f6413c7b391e01eec779a701ff16f0b15f10f459b963ef06" exitCode=0 Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.223945 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ff8j2" event={"ID":"2408c6a0-ae42-4da0-a8d8-262db680df45","Type":"ContainerDied","Data":"e126d8f8270a0862f6413c7b391e01eec779a701ff16f0b15f10f459b963ef06"} Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.223993 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ff8j2" event={"ID":"2408c6a0-ae42-4da0-a8d8-262db680df45","Type":"ContainerDied","Data":"b7541a7b4d560940e36387a4af2f85404c194e647b15a386706a39d16411a535"} Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.224008 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ff8j2" Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.224016 4809 scope.go:117] "RemoveContainer" containerID="e126d8f8270a0862f6413c7b391e01eec779a701ff16f0b15f10f459b963ef06" Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.248107 4809 scope.go:117] "RemoveContainer" containerID="fb6839ee6d8f8b5f210bc7968ac75e9c84d5ed0566a51c5fee08e3ed173dd9e7" Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.257737 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ff8j2"] Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.266462 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ff8j2"] Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.271067 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2408c6a0-ae42-4da0-a8d8-262db680df45-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.271094 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzsjb\" (UniqueName: \"kubernetes.io/projected/2408c6a0-ae42-4da0-a8d8-262db680df45-kube-api-access-xzsjb\") on node \"crc\" DevicePath \"\"" Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.271103 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2408c6a0-ae42-4da0-a8d8-262db680df45-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.277141 4809 scope.go:117] "RemoveContainer" containerID="ac2aef5b94e892d15507bd4855919bb279fd5dcdbfd4f519563ff7c49faf4cea" Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.318334 4809 scope.go:117] "RemoveContainer" containerID="e126d8f8270a0862f6413c7b391e01eec779a701ff16f0b15f10f459b963ef06" Nov 24 07:42:41 crc kubenswrapper[4809]: E1124 07:42:41.319005 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e126d8f8270a0862f6413c7b391e01eec779a701ff16f0b15f10f459b963ef06\": container with ID starting with e126d8f8270a0862f6413c7b391e01eec779a701ff16f0b15f10f459b963ef06 not found: ID does not exist" containerID="e126d8f8270a0862f6413c7b391e01eec779a701ff16f0b15f10f459b963ef06" Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.319045 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e126d8f8270a0862f6413c7b391e01eec779a701ff16f0b15f10f459b963ef06"} err="failed to get container status \"e126d8f8270a0862f6413c7b391e01eec779a701ff16f0b15f10f459b963ef06\": rpc error: code = NotFound desc = could not find container \"e126d8f8270a0862f6413c7b391e01eec779a701ff16f0b15f10f459b963ef06\": container with ID starting with e126d8f8270a0862f6413c7b391e01eec779a701ff16f0b15f10f459b963ef06 not found: ID does not exist" Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.319071 4809 scope.go:117] "RemoveContainer" containerID="fb6839ee6d8f8b5f210bc7968ac75e9c84d5ed0566a51c5fee08e3ed173dd9e7" Nov 24 07:42:41 crc kubenswrapper[4809]: E1124 07:42:41.319396 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb6839ee6d8f8b5f210bc7968ac75e9c84d5ed0566a51c5fee08e3ed173dd9e7\": container with ID starting with fb6839ee6d8f8b5f210bc7968ac75e9c84d5ed0566a51c5fee08e3ed173dd9e7 not found: ID does not exist" containerID="fb6839ee6d8f8b5f210bc7968ac75e9c84d5ed0566a51c5fee08e3ed173dd9e7" Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.319418 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb6839ee6d8f8b5f210bc7968ac75e9c84d5ed0566a51c5fee08e3ed173dd9e7"} err="failed to get container status \"fb6839ee6d8f8b5f210bc7968ac75e9c84d5ed0566a51c5fee08e3ed173dd9e7\": rpc error: code = NotFound desc = could not find container \"fb6839ee6d8f8b5f210bc7968ac75e9c84d5ed0566a51c5fee08e3ed173dd9e7\": container with ID starting with fb6839ee6d8f8b5f210bc7968ac75e9c84d5ed0566a51c5fee08e3ed173dd9e7 not found: ID does not exist" Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.319431 4809 scope.go:117] "RemoveContainer" containerID="ac2aef5b94e892d15507bd4855919bb279fd5dcdbfd4f519563ff7c49faf4cea" Nov 24 07:42:41 crc kubenswrapper[4809]: E1124 07:42:41.319723 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac2aef5b94e892d15507bd4855919bb279fd5dcdbfd4f519563ff7c49faf4cea\": container with ID starting with ac2aef5b94e892d15507bd4855919bb279fd5dcdbfd4f519563ff7c49faf4cea not found: ID does not exist" containerID="ac2aef5b94e892d15507bd4855919bb279fd5dcdbfd4f519563ff7c49faf4cea" Nov 24 07:42:41 crc kubenswrapper[4809]: I1124 07:42:41.319745 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac2aef5b94e892d15507bd4855919bb279fd5dcdbfd4f519563ff7c49faf4cea"} err="failed to get container status \"ac2aef5b94e892d15507bd4855919bb279fd5dcdbfd4f519563ff7c49faf4cea\": rpc error: code = NotFound desc = could not find container \"ac2aef5b94e892d15507bd4855919bb279fd5dcdbfd4f519563ff7c49faf4cea\": container with ID starting with ac2aef5b94e892d15507bd4855919bb279fd5dcdbfd4f519563ff7c49faf4cea not found: ID does not exist" Nov 24 07:42:42 crc kubenswrapper[4809]: I1124 07:42:42.900807 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2408c6a0-ae42-4da0-a8d8-262db680df45" path="/var/lib/kubelet/pods/2408c6a0-ae42-4da0-a8d8-262db680df45/volumes" Nov 24 07:42:48 crc kubenswrapper[4809]: I1124 07:42:48.043036 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:42:48 crc kubenswrapper[4809]: I1124 07:42:48.043859 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:43:18 crc kubenswrapper[4809]: I1124 07:43:18.043379 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:43:18 crc kubenswrapper[4809]: I1124 07:43:18.043835 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:43:18 crc kubenswrapper[4809]: I1124 07:43:18.043877 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 07:43:18 crc kubenswrapper[4809]: I1124 07:43:18.044564 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759"} pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:43:18 crc kubenswrapper[4809]: I1124 07:43:18.044618 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" containerID="cri-o://1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" gracePeriod=600 Nov 24 07:43:18 crc kubenswrapper[4809]: E1124 07:43:18.175908 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:43:18 crc kubenswrapper[4809]: I1124 07:43:18.573152 4809 generic.go:334] "Generic (PLEG): container finished" podID="be8050d5-1aea-453e-a06e-87702b0e856b" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" exitCode=0 Nov 24 07:43:18 crc kubenswrapper[4809]: I1124 07:43:18.573203 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerDied","Data":"1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759"} Nov 24 07:43:18 crc kubenswrapper[4809]: I1124 07:43:18.573241 4809 scope.go:117] "RemoveContainer" containerID="630f55bb399f3bfd04e8309b3cf29adad95cffa6451d4ba31d28cc6a268df881" Nov 24 07:43:18 crc kubenswrapper[4809]: I1124 07:43:18.577766 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:43:18 crc kubenswrapper[4809]: E1124 07:43:18.578315 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:43:33 crc kubenswrapper[4809]: I1124 07:43:33.891067 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:43:33 crc kubenswrapper[4809]: E1124 07:43:33.891939 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:43:44 crc kubenswrapper[4809]: I1124 07:43:44.891935 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:43:44 crc kubenswrapper[4809]: E1124 07:43:44.892741 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:43:57 crc kubenswrapper[4809]: I1124 07:43:57.892022 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:43:57 crc kubenswrapper[4809]: E1124 07:43:57.893249 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:44:10 crc kubenswrapper[4809]: I1124 07:44:10.908342 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:44:10 crc kubenswrapper[4809]: E1124 07:44:10.909161 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:44:24 crc kubenswrapper[4809]: I1124 07:44:24.891360 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:44:24 crc kubenswrapper[4809]: E1124 07:44:24.892049 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:44:37 crc kubenswrapper[4809]: I1124 07:44:37.891806 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:44:37 crc kubenswrapper[4809]: E1124 07:44:37.892911 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:44:51 crc kubenswrapper[4809]: I1124 07:44:51.891685 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:44:51 crc kubenswrapper[4809]: E1124 07:44:51.892357 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:45:00 crc kubenswrapper[4809]: I1124 07:45:00.173583 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399505-zb5q6"] Nov 24 07:45:00 crc kubenswrapper[4809]: E1124 07:45:00.174604 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2408c6a0-ae42-4da0-a8d8-262db680df45" containerName="extract-utilities" Nov 24 07:45:00 crc kubenswrapper[4809]: I1124 07:45:00.174623 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2408c6a0-ae42-4da0-a8d8-262db680df45" containerName="extract-utilities" Nov 24 07:45:00 crc kubenswrapper[4809]: E1124 07:45:00.174641 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2408c6a0-ae42-4da0-a8d8-262db680df45" containerName="registry-server" Nov 24 07:45:00 crc kubenswrapper[4809]: I1124 07:45:00.174649 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2408c6a0-ae42-4da0-a8d8-262db680df45" containerName="registry-server" Nov 24 07:45:00 crc kubenswrapper[4809]: E1124 07:45:00.174664 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2408c6a0-ae42-4da0-a8d8-262db680df45" containerName="extract-content" Nov 24 07:45:00 crc kubenswrapper[4809]: I1124 07:45:00.174672 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2408c6a0-ae42-4da0-a8d8-262db680df45" containerName="extract-content" Nov 24 07:45:00 crc kubenswrapper[4809]: I1124 07:45:00.174936 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2408c6a0-ae42-4da0-a8d8-262db680df45" containerName="registry-server" Nov 24 07:45:00 crc kubenswrapper[4809]: I1124 07:45:00.176085 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-zb5q6" Nov 24 07:45:00 crc kubenswrapper[4809]: I1124 07:45:00.182239 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 07:45:00 crc kubenswrapper[4809]: I1124 07:45:00.183750 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 07:45:00 crc kubenswrapper[4809]: I1124 07:45:00.186899 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399505-zb5q6"] Nov 24 07:45:00 crc kubenswrapper[4809]: I1124 07:45:00.336696 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2c0d677-b391-4958-8439-06e31a2074e3-secret-volume\") pod \"collect-profiles-29399505-zb5q6\" (UID: \"f2c0d677-b391-4958-8439-06e31a2074e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-zb5q6" Nov 24 07:45:00 crc kubenswrapper[4809]: I1124 07:45:00.337149 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqbd2\" (UniqueName: \"kubernetes.io/projected/f2c0d677-b391-4958-8439-06e31a2074e3-kube-api-access-zqbd2\") pod \"collect-profiles-29399505-zb5q6\" (UID: \"f2c0d677-b391-4958-8439-06e31a2074e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-zb5q6" Nov 24 07:45:00 crc kubenswrapper[4809]: I1124 07:45:00.337257 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2c0d677-b391-4958-8439-06e31a2074e3-config-volume\") pod \"collect-profiles-29399505-zb5q6\" (UID: \"f2c0d677-b391-4958-8439-06e31a2074e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-zb5q6" Nov 24 07:45:00 crc kubenswrapper[4809]: I1124 07:45:00.438515 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqbd2\" (UniqueName: \"kubernetes.io/projected/f2c0d677-b391-4958-8439-06e31a2074e3-kube-api-access-zqbd2\") pod \"collect-profiles-29399505-zb5q6\" (UID: \"f2c0d677-b391-4958-8439-06e31a2074e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-zb5q6" Nov 24 07:45:00 crc kubenswrapper[4809]: I1124 07:45:00.438577 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2c0d677-b391-4958-8439-06e31a2074e3-config-volume\") pod \"collect-profiles-29399505-zb5q6\" (UID: \"f2c0d677-b391-4958-8439-06e31a2074e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-zb5q6" Nov 24 07:45:00 crc kubenswrapper[4809]: I1124 07:45:00.438686 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2c0d677-b391-4958-8439-06e31a2074e3-secret-volume\") pod \"collect-profiles-29399505-zb5q6\" (UID: \"f2c0d677-b391-4958-8439-06e31a2074e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-zb5q6" Nov 24 07:45:00 crc kubenswrapper[4809]: I1124 07:45:00.439671 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2c0d677-b391-4958-8439-06e31a2074e3-config-volume\") pod \"collect-profiles-29399505-zb5q6\" (UID: \"f2c0d677-b391-4958-8439-06e31a2074e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-zb5q6" Nov 24 07:45:00 crc kubenswrapper[4809]: I1124 07:45:00.444760 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2c0d677-b391-4958-8439-06e31a2074e3-secret-volume\") pod \"collect-profiles-29399505-zb5q6\" (UID: \"f2c0d677-b391-4958-8439-06e31a2074e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-zb5q6" Nov 24 07:45:00 crc kubenswrapper[4809]: I1124 07:45:00.453207 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqbd2\" (UniqueName: \"kubernetes.io/projected/f2c0d677-b391-4958-8439-06e31a2074e3-kube-api-access-zqbd2\") pod \"collect-profiles-29399505-zb5q6\" (UID: \"f2c0d677-b391-4958-8439-06e31a2074e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-zb5q6" Nov 24 07:45:00 crc kubenswrapper[4809]: I1124 07:45:00.501136 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-zb5q6" Nov 24 07:45:00 crc kubenswrapper[4809]: I1124 07:45:00.922661 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399505-zb5q6"] Nov 24 07:45:01 crc kubenswrapper[4809]: I1124 07:45:01.482306 4809 generic.go:334] "Generic (PLEG): container finished" podID="f2c0d677-b391-4958-8439-06e31a2074e3" containerID="fd713a1e869c007afa3efd5561c2f2811ad7a8b1f57a0268502f00f0cb739d00" exitCode=0 Nov 24 07:45:01 crc kubenswrapper[4809]: I1124 07:45:01.482363 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-zb5q6" event={"ID":"f2c0d677-b391-4958-8439-06e31a2074e3","Type":"ContainerDied","Data":"fd713a1e869c007afa3efd5561c2f2811ad7a8b1f57a0268502f00f0cb739d00"} Nov 24 07:45:01 crc kubenswrapper[4809]: I1124 07:45:01.482641 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-zb5q6" event={"ID":"f2c0d677-b391-4958-8439-06e31a2074e3","Type":"ContainerStarted","Data":"1ef9ed09e4072a945d0cd3dff3d8c491036e748b6ac87c36f228d58db585070a"} Nov 24 07:45:02 crc kubenswrapper[4809]: I1124 07:45:02.844794 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-zb5q6" Nov 24 07:45:02 crc kubenswrapper[4809]: I1124 07:45:02.986595 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqbd2\" (UniqueName: \"kubernetes.io/projected/f2c0d677-b391-4958-8439-06e31a2074e3-kube-api-access-zqbd2\") pod \"f2c0d677-b391-4958-8439-06e31a2074e3\" (UID: \"f2c0d677-b391-4958-8439-06e31a2074e3\") " Nov 24 07:45:02 crc kubenswrapper[4809]: I1124 07:45:02.987089 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2c0d677-b391-4958-8439-06e31a2074e3-config-volume\") pod \"f2c0d677-b391-4958-8439-06e31a2074e3\" (UID: \"f2c0d677-b391-4958-8439-06e31a2074e3\") " Nov 24 07:45:02 crc kubenswrapper[4809]: I1124 07:45:02.987255 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2c0d677-b391-4958-8439-06e31a2074e3-secret-volume\") pod \"f2c0d677-b391-4958-8439-06e31a2074e3\" (UID: \"f2c0d677-b391-4958-8439-06e31a2074e3\") " Nov 24 07:45:02 crc kubenswrapper[4809]: I1124 07:45:02.987798 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2c0d677-b391-4958-8439-06e31a2074e3-config-volume" (OuterVolumeSpecName: "config-volume") pod "f2c0d677-b391-4958-8439-06e31a2074e3" (UID: "f2c0d677-b391-4958-8439-06e31a2074e3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:45:02 crc kubenswrapper[4809]: I1124 07:45:02.991872 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2c0d677-b391-4958-8439-06e31a2074e3-kube-api-access-zqbd2" (OuterVolumeSpecName: "kube-api-access-zqbd2") pod "f2c0d677-b391-4958-8439-06e31a2074e3" (UID: "f2c0d677-b391-4958-8439-06e31a2074e3"). InnerVolumeSpecName "kube-api-access-zqbd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:45:02 crc kubenswrapper[4809]: I1124 07:45:02.994034 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2c0d677-b391-4958-8439-06e31a2074e3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f2c0d677-b391-4958-8439-06e31a2074e3" (UID: "f2c0d677-b391-4958-8439-06e31a2074e3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:45:03 crc kubenswrapper[4809]: I1124 07:45:03.089803 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2c0d677-b391-4958-8439-06e31a2074e3-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 07:45:03 crc kubenswrapper[4809]: I1124 07:45:03.089834 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2c0d677-b391-4958-8439-06e31a2074e3-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 07:45:03 crc kubenswrapper[4809]: I1124 07:45:03.089843 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqbd2\" (UniqueName: \"kubernetes.io/projected/f2c0d677-b391-4958-8439-06e31a2074e3-kube-api-access-zqbd2\") on node \"crc\" DevicePath \"\"" Nov 24 07:45:03 crc kubenswrapper[4809]: I1124 07:45:03.499641 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-zb5q6" event={"ID":"f2c0d677-b391-4958-8439-06e31a2074e3","Type":"ContainerDied","Data":"1ef9ed09e4072a945d0cd3dff3d8c491036e748b6ac87c36f228d58db585070a"} Nov 24 07:45:03 crc kubenswrapper[4809]: I1124 07:45:03.499689 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-zb5q6" Nov 24 07:45:03 crc kubenswrapper[4809]: I1124 07:45:03.499698 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ef9ed09e4072a945d0cd3dff3d8c491036e748b6ac87c36f228d58db585070a" Nov 24 07:45:03 crc kubenswrapper[4809]: I1124 07:45:03.918931 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399460-hdrql"] Nov 24 07:45:03 crc kubenswrapper[4809]: I1124 07:45:03.929137 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399460-hdrql"] Nov 24 07:45:04 crc kubenswrapper[4809]: I1124 07:45:04.903623 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f2df143-a48b-46eb-9b59-3903504fe5a4" path="/var/lib/kubelet/pods/2f2df143-a48b-46eb-9b59-3903504fe5a4/volumes" Nov 24 07:45:06 crc kubenswrapper[4809]: I1124 07:45:06.893195 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:45:06 crc kubenswrapper[4809]: E1124 07:45:06.893960 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:45:10 crc kubenswrapper[4809]: I1124 07:45:10.500620 4809 scope.go:117] "RemoveContainer" containerID="636cfddffffee7ce696a864d7631a31103588c2f5d13834a00a576d89f6e95b4" Nov 24 07:45:18 crc kubenswrapper[4809]: I1124 07:45:18.890722 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:45:18 crc kubenswrapper[4809]: E1124 07:45:18.891497 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:45:33 crc kubenswrapper[4809]: I1124 07:45:33.891017 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:45:33 crc kubenswrapper[4809]: E1124 07:45:33.891878 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:45:47 crc kubenswrapper[4809]: I1124 07:45:47.891720 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:45:47 crc kubenswrapper[4809]: E1124 07:45:47.892508 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:46:00 crc kubenswrapper[4809]: I1124 07:46:00.899943 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:46:00 crc kubenswrapper[4809]: E1124 07:46:00.900833 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:46:11 crc kubenswrapper[4809]: I1124 07:46:11.890821 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:46:11 crc kubenswrapper[4809]: E1124 07:46:11.891701 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:46:22 crc kubenswrapper[4809]: I1124 07:46:22.891327 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:46:22 crc kubenswrapper[4809]: E1124 07:46:22.892072 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:46:34 crc kubenswrapper[4809]: I1124 07:46:34.892572 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:46:34 crc kubenswrapper[4809]: E1124 07:46:34.893338 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:46:45 crc kubenswrapper[4809]: I1124 07:46:45.891676 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:46:45 crc kubenswrapper[4809]: E1124 07:46:45.892458 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:47:00 crc kubenswrapper[4809]: I1124 07:47:00.911505 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:47:00 crc kubenswrapper[4809]: E1124 07:47:00.912653 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:47:12 crc kubenswrapper[4809]: I1124 07:47:12.892170 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:47:12 crc kubenswrapper[4809]: E1124 07:47:12.893312 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:47:23 crc kubenswrapper[4809]: I1124 07:47:23.903841 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:47:23 crc kubenswrapper[4809]: E1124 07:47:23.905423 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:47:38 crc kubenswrapper[4809]: I1124 07:47:38.892246 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:47:38 crc kubenswrapper[4809]: E1124 07:47:38.893000 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:47:52 crc kubenswrapper[4809]: I1124 07:47:52.891656 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:47:52 crc kubenswrapper[4809]: E1124 07:47:52.892498 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:48:04 crc kubenswrapper[4809]: I1124 07:48:04.892202 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:48:04 crc kubenswrapper[4809]: E1124 07:48:04.894283 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:48:15 crc kubenswrapper[4809]: I1124 07:48:15.891081 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:48:15 crc kubenswrapper[4809]: E1124 07:48:15.891763 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:48:26 crc kubenswrapper[4809]: I1124 07:48:26.891818 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:48:27 crc kubenswrapper[4809]: I1124 07:48:27.387412 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerStarted","Data":"7d5c0ef1309a4cda8c3a374bbff14973c66d06a10b1ee334b7a60d88f824f490"} Nov 24 07:50:08 crc kubenswrapper[4809]: I1124 07:50:08.924012 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xlbb6"] Nov 24 07:50:08 crc kubenswrapper[4809]: E1124 07:50:08.925184 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2c0d677-b391-4958-8439-06e31a2074e3" containerName="collect-profiles" Nov 24 07:50:08 crc kubenswrapper[4809]: I1124 07:50:08.925201 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2c0d677-b391-4958-8439-06e31a2074e3" containerName="collect-profiles" Nov 24 07:50:08 crc kubenswrapper[4809]: I1124 07:50:08.925420 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2c0d677-b391-4958-8439-06e31a2074e3" containerName="collect-profiles" Nov 24 07:50:08 crc kubenswrapper[4809]: I1124 07:50:08.932573 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xlbb6" Nov 24 07:50:08 crc kubenswrapper[4809]: I1124 07:50:08.960950 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xlbb6"] Nov 24 07:50:09 crc kubenswrapper[4809]: I1124 07:50:09.069838 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44b3dee8-179f-486a-8138-c419c2fead38-catalog-content\") pod \"redhat-operators-xlbb6\" (UID: \"44b3dee8-179f-486a-8138-c419c2fead38\") " pod="openshift-marketplace/redhat-operators-xlbb6" Nov 24 07:50:09 crc kubenswrapper[4809]: I1124 07:50:09.069913 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44b3dee8-179f-486a-8138-c419c2fead38-utilities\") pod \"redhat-operators-xlbb6\" (UID: \"44b3dee8-179f-486a-8138-c419c2fead38\") " pod="openshift-marketplace/redhat-operators-xlbb6" Nov 24 07:50:09 crc kubenswrapper[4809]: I1124 07:50:09.070046 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kspm\" (UniqueName: \"kubernetes.io/projected/44b3dee8-179f-486a-8138-c419c2fead38-kube-api-access-7kspm\") pod \"redhat-operators-xlbb6\" (UID: \"44b3dee8-179f-486a-8138-c419c2fead38\") " pod="openshift-marketplace/redhat-operators-xlbb6" Nov 24 07:50:09 crc kubenswrapper[4809]: I1124 07:50:09.172352 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44b3dee8-179f-486a-8138-c419c2fead38-catalog-content\") pod \"redhat-operators-xlbb6\" (UID: \"44b3dee8-179f-486a-8138-c419c2fead38\") " pod="openshift-marketplace/redhat-operators-xlbb6" Nov 24 07:50:09 crc kubenswrapper[4809]: I1124 07:50:09.172423 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44b3dee8-179f-486a-8138-c419c2fead38-utilities\") pod \"redhat-operators-xlbb6\" (UID: \"44b3dee8-179f-486a-8138-c419c2fead38\") " pod="openshift-marketplace/redhat-operators-xlbb6" Nov 24 07:50:09 crc kubenswrapper[4809]: I1124 07:50:09.172540 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kspm\" (UniqueName: \"kubernetes.io/projected/44b3dee8-179f-486a-8138-c419c2fead38-kube-api-access-7kspm\") pod \"redhat-operators-xlbb6\" (UID: \"44b3dee8-179f-486a-8138-c419c2fead38\") " pod="openshift-marketplace/redhat-operators-xlbb6" Nov 24 07:50:09 crc kubenswrapper[4809]: I1124 07:50:09.172934 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44b3dee8-179f-486a-8138-c419c2fead38-catalog-content\") pod \"redhat-operators-xlbb6\" (UID: \"44b3dee8-179f-486a-8138-c419c2fead38\") " pod="openshift-marketplace/redhat-operators-xlbb6" Nov 24 07:50:09 crc kubenswrapper[4809]: I1124 07:50:09.173026 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44b3dee8-179f-486a-8138-c419c2fead38-utilities\") pod \"redhat-operators-xlbb6\" (UID: \"44b3dee8-179f-486a-8138-c419c2fead38\") " pod="openshift-marketplace/redhat-operators-xlbb6" Nov 24 07:50:09 crc kubenswrapper[4809]: I1124 07:50:09.193112 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kspm\" (UniqueName: \"kubernetes.io/projected/44b3dee8-179f-486a-8138-c419c2fead38-kube-api-access-7kspm\") pod \"redhat-operators-xlbb6\" (UID: \"44b3dee8-179f-486a-8138-c419c2fead38\") " pod="openshift-marketplace/redhat-operators-xlbb6" Nov 24 07:50:09 crc kubenswrapper[4809]: I1124 07:50:09.270913 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xlbb6" Nov 24 07:50:09 crc kubenswrapper[4809]: I1124 07:50:09.756895 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xlbb6"] Nov 24 07:50:10 crc kubenswrapper[4809]: I1124 07:50:10.284275 4809 generic.go:334] "Generic (PLEG): container finished" podID="44b3dee8-179f-486a-8138-c419c2fead38" containerID="2d78d16c996ad2a71d6ae647bd53162d9354be184a52bafab845db02fc7003f4" exitCode=0 Nov 24 07:50:10 crc kubenswrapper[4809]: I1124 07:50:10.284329 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlbb6" event={"ID":"44b3dee8-179f-486a-8138-c419c2fead38","Type":"ContainerDied","Data":"2d78d16c996ad2a71d6ae647bd53162d9354be184a52bafab845db02fc7003f4"} Nov 24 07:50:10 crc kubenswrapper[4809]: I1124 07:50:10.284362 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlbb6" event={"ID":"44b3dee8-179f-486a-8138-c419c2fead38","Type":"ContainerStarted","Data":"f97967c7cdc70bf2765aa1704e4418e6207e6f0e262997035c455bf65227b5fb"} Nov 24 07:50:10 crc kubenswrapper[4809]: I1124 07:50:10.286546 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 07:50:12 crc kubenswrapper[4809]: I1124 07:50:12.302767 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlbb6" event={"ID":"44b3dee8-179f-486a-8138-c419c2fead38","Type":"ContainerStarted","Data":"a10396bae4fd7fde9cf7f2e97a0ed26ac6063d497cf5b109c83e985ae835c21e"} Nov 24 07:50:15 crc kubenswrapper[4809]: I1124 07:50:15.339381 4809 generic.go:334] "Generic (PLEG): container finished" podID="44b3dee8-179f-486a-8138-c419c2fead38" containerID="a10396bae4fd7fde9cf7f2e97a0ed26ac6063d497cf5b109c83e985ae835c21e" exitCode=0 Nov 24 07:50:15 crc kubenswrapper[4809]: I1124 07:50:15.339513 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlbb6" event={"ID":"44b3dee8-179f-486a-8138-c419c2fead38","Type":"ContainerDied","Data":"a10396bae4fd7fde9cf7f2e97a0ed26ac6063d497cf5b109c83e985ae835c21e"} Nov 24 07:50:16 crc kubenswrapper[4809]: I1124 07:50:16.363213 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlbb6" event={"ID":"44b3dee8-179f-486a-8138-c419c2fead38","Type":"ContainerStarted","Data":"1d1193c07b790cfae0ec63b3f92d103869a4bf9bdf91cd2c39c48405a591a709"} Nov 24 07:50:16 crc kubenswrapper[4809]: I1124 07:50:16.391316 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xlbb6" podStartSLOduration=2.913171779 podStartE2EDuration="8.391296251s" podCreationTimestamp="2025-11-24 07:50:08 +0000 UTC" firstStartedPulling="2025-11-24 07:50:10.286334685 +0000 UTC m=+3310.186926390" lastFinishedPulling="2025-11-24 07:50:15.764459157 +0000 UTC m=+3315.665050862" observedRunningTime="2025-11-24 07:50:16.389345438 +0000 UTC m=+3316.289937143" watchObservedRunningTime="2025-11-24 07:50:16.391296251 +0000 UTC m=+3316.291887956" Nov 24 07:50:19 crc kubenswrapper[4809]: I1124 07:50:19.271742 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xlbb6" Nov 24 07:50:19 crc kubenswrapper[4809]: I1124 07:50:19.272152 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xlbb6" Nov 24 07:50:20 crc kubenswrapper[4809]: I1124 07:50:20.315870 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xlbb6" podUID="44b3dee8-179f-486a-8138-c419c2fead38" containerName="registry-server" probeResult="failure" output=< Nov 24 07:50:20 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Nov 24 07:50:20 crc kubenswrapper[4809]: > Nov 24 07:50:25 crc kubenswrapper[4809]: I1124 07:50:25.436434 4809 generic.go:334] "Generic (PLEG): container finished" podID="4f90573b-06dc-47d5-ab8b-5dab5503e96a" containerID="b3eedb3a6223893f8c6be0a1b50d68784be92462929ec43f020a7f4833601c88" exitCode=0 Nov 24 07:50:25 crc kubenswrapper[4809]: I1124 07:50:25.436629 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"4f90573b-06dc-47d5-ab8b-5dab5503e96a","Type":"ContainerDied","Data":"b3eedb3a6223893f8c6be0a1b50d68784be92462929ec43f020a7f4833601c88"} Nov 24 07:50:26 crc kubenswrapper[4809]: I1124 07:50:26.874511 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.000019 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f90573b-06dc-47d5-ab8b-5dab5503e96a-ssh-key\") pod \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.000058 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.000094 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4f90573b-06dc-47d5-ab8b-5dab5503e96a-openstack-config-secret\") pod \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.000124 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tb6vh\" (UniqueName: \"kubernetes.io/projected/4f90573b-06dc-47d5-ab8b-5dab5503e96a-kube-api-access-tb6vh\") pod \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.000218 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4f90573b-06dc-47d5-ab8b-5dab5503e96a-ca-certs\") pod \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.000242 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4f90573b-06dc-47d5-ab8b-5dab5503e96a-config-data\") pod \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.000258 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4f90573b-06dc-47d5-ab8b-5dab5503e96a-openstack-config\") pod \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.000277 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4f90573b-06dc-47d5-ab8b-5dab5503e96a-test-operator-ephemeral-workdir\") pod \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.000315 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4f90573b-06dc-47d5-ab8b-5dab5503e96a-test-operator-ephemeral-temporary\") pod \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\" (UID: \"4f90573b-06dc-47d5-ab8b-5dab5503e96a\") " Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.001163 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f90573b-06dc-47d5-ab8b-5dab5503e96a-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "4f90573b-06dc-47d5-ab8b-5dab5503e96a" (UID: "4f90573b-06dc-47d5-ab8b-5dab5503e96a"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.001229 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f90573b-06dc-47d5-ab8b-5dab5503e96a-config-data" (OuterVolumeSpecName: "config-data") pod "4f90573b-06dc-47d5-ab8b-5dab5503e96a" (UID: "4f90573b-06dc-47d5-ab8b-5dab5503e96a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.005606 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f90573b-06dc-47d5-ab8b-5dab5503e96a-kube-api-access-tb6vh" (OuterVolumeSpecName: "kube-api-access-tb6vh") pod "4f90573b-06dc-47d5-ab8b-5dab5503e96a" (UID: "4f90573b-06dc-47d5-ab8b-5dab5503e96a"). InnerVolumeSpecName "kube-api-access-tb6vh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.005916 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f90573b-06dc-47d5-ab8b-5dab5503e96a-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "4f90573b-06dc-47d5-ab8b-5dab5503e96a" (UID: "4f90573b-06dc-47d5-ab8b-5dab5503e96a"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.006227 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "test-operator-logs") pod "4f90573b-06dc-47d5-ab8b-5dab5503e96a" (UID: "4f90573b-06dc-47d5-ab8b-5dab5503e96a"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.006743 4809 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4f90573b-06dc-47d5-ab8b-5dab5503e96a-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.006789 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.006802 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tb6vh\" (UniqueName: \"kubernetes.io/projected/4f90573b-06dc-47d5-ab8b-5dab5503e96a-kube-api-access-tb6vh\") on node \"crc\" DevicePath \"\"" Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.006813 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4f90573b-06dc-47d5-ab8b-5dab5503e96a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.006823 4809 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4f90573b-06dc-47d5-ab8b-5dab5503e96a-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.029466 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f90573b-06dc-47d5-ab8b-5dab5503e96a-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "4f90573b-06dc-47d5-ab8b-5dab5503e96a" (UID: "4f90573b-06dc-47d5-ab8b-5dab5503e96a"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.031254 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.032778 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f90573b-06dc-47d5-ab8b-5dab5503e96a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4f90573b-06dc-47d5-ab8b-5dab5503e96a" (UID: "4f90573b-06dc-47d5-ab8b-5dab5503e96a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.036216 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f90573b-06dc-47d5-ab8b-5dab5503e96a-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "4f90573b-06dc-47d5-ab8b-5dab5503e96a" (UID: "4f90573b-06dc-47d5-ab8b-5dab5503e96a"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.056455 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f90573b-06dc-47d5-ab8b-5dab5503e96a-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "4f90573b-06dc-47d5-ab8b-5dab5503e96a" (UID: "4f90573b-06dc-47d5-ab8b-5dab5503e96a"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.108735 4809 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4f90573b-06dc-47d5-ab8b-5dab5503e96a-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.108767 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4f90573b-06dc-47d5-ab8b-5dab5503e96a-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.108805 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f90573b-06dc-47d5-ab8b-5dab5503e96a-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.108814 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.108823 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4f90573b-06dc-47d5-ab8b-5dab5503e96a-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.457285 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"4f90573b-06dc-47d5-ab8b-5dab5503e96a","Type":"ContainerDied","Data":"47000287f8bcad8ad972a79926af841302b45adb3f916bf0c837d7725cdc7551"} Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.457328 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47000287f8bcad8ad972a79926af841302b45adb3f916bf0c837d7725cdc7551" Nov 24 07:50:27 crc kubenswrapper[4809]: I1124 07:50:27.457363 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 07:50:29 crc kubenswrapper[4809]: I1124 07:50:29.323374 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xlbb6" Nov 24 07:50:29 crc kubenswrapper[4809]: I1124 07:50:29.400847 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xlbb6" Nov 24 07:50:29 crc kubenswrapper[4809]: I1124 07:50:29.409248 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 07:50:29 crc kubenswrapper[4809]: E1124 07:50:29.409791 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f90573b-06dc-47d5-ab8b-5dab5503e96a" containerName="tempest-tests-tempest-tests-runner" Nov 24 07:50:29 crc kubenswrapper[4809]: I1124 07:50:29.409813 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f90573b-06dc-47d5-ab8b-5dab5503e96a" containerName="tempest-tests-tempest-tests-runner" Nov 24 07:50:29 crc kubenswrapper[4809]: I1124 07:50:29.410099 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f90573b-06dc-47d5-ab8b-5dab5503e96a" containerName="tempest-tests-tempest-tests-runner" Nov 24 07:50:29 crc kubenswrapper[4809]: I1124 07:50:29.410903 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 07:50:29 crc kubenswrapper[4809]: I1124 07:50:29.414252 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-dqvf8" Nov 24 07:50:29 crc kubenswrapper[4809]: I1124 07:50:29.425177 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 07:50:29 crc kubenswrapper[4809]: I1124 07:50:29.562375 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xlbb6"] Nov 24 07:50:29 crc kubenswrapper[4809]: I1124 07:50:29.565074 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v58d2\" (UniqueName: \"kubernetes.io/projected/19e09f12-9159-4150-ae5b-6d718e6f9957-kube-api-access-v58d2\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"19e09f12-9159-4150-ae5b-6d718e6f9957\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 07:50:29 crc kubenswrapper[4809]: I1124 07:50:29.565168 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"19e09f12-9159-4150-ae5b-6d718e6f9957\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 07:50:29 crc kubenswrapper[4809]: I1124 07:50:29.666922 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"19e09f12-9159-4150-ae5b-6d718e6f9957\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 07:50:29 crc kubenswrapper[4809]: I1124 07:50:29.667298 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v58d2\" (UniqueName: \"kubernetes.io/projected/19e09f12-9159-4150-ae5b-6d718e6f9957-kube-api-access-v58d2\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"19e09f12-9159-4150-ae5b-6d718e6f9957\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 07:50:29 crc kubenswrapper[4809]: I1124 07:50:29.667419 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"19e09f12-9159-4150-ae5b-6d718e6f9957\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 07:50:29 crc kubenswrapper[4809]: I1124 07:50:29.693755 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v58d2\" (UniqueName: \"kubernetes.io/projected/19e09f12-9159-4150-ae5b-6d718e6f9957-kube-api-access-v58d2\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"19e09f12-9159-4150-ae5b-6d718e6f9957\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 07:50:29 crc kubenswrapper[4809]: I1124 07:50:29.699917 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"19e09f12-9159-4150-ae5b-6d718e6f9957\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 07:50:29 crc kubenswrapper[4809]: I1124 07:50:29.739614 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 07:50:30 crc kubenswrapper[4809]: I1124 07:50:30.200450 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 07:50:30 crc kubenswrapper[4809]: I1124 07:50:30.492289 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xlbb6" podUID="44b3dee8-179f-486a-8138-c419c2fead38" containerName="registry-server" containerID="cri-o://1d1193c07b790cfae0ec63b3f92d103869a4bf9bdf91cd2c39c48405a591a709" gracePeriod=2 Nov 24 07:50:30 crc kubenswrapper[4809]: I1124 07:50:30.493095 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"19e09f12-9159-4150-ae5b-6d718e6f9957","Type":"ContainerStarted","Data":"5ecec483337a4022ee0dedd668dd07d1846021c7604f3ec4146204c1ea7b4f61"} Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.165114 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xlbb6" Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.304202 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44b3dee8-179f-486a-8138-c419c2fead38-catalog-content\") pod \"44b3dee8-179f-486a-8138-c419c2fead38\" (UID: \"44b3dee8-179f-486a-8138-c419c2fead38\") " Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.304543 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kspm\" (UniqueName: \"kubernetes.io/projected/44b3dee8-179f-486a-8138-c419c2fead38-kube-api-access-7kspm\") pod \"44b3dee8-179f-486a-8138-c419c2fead38\" (UID: \"44b3dee8-179f-486a-8138-c419c2fead38\") " Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.304740 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44b3dee8-179f-486a-8138-c419c2fead38-utilities\") pod \"44b3dee8-179f-486a-8138-c419c2fead38\" (UID: \"44b3dee8-179f-486a-8138-c419c2fead38\") " Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.305808 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44b3dee8-179f-486a-8138-c419c2fead38-utilities" (OuterVolumeSpecName: "utilities") pod "44b3dee8-179f-486a-8138-c419c2fead38" (UID: "44b3dee8-179f-486a-8138-c419c2fead38"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.313182 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44b3dee8-179f-486a-8138-c419c2fead38-kube-api-access-7kspm" (OuterVolumeSpecName: "kube-api-access-7kspm") pod "44b3dee8-179f-486a-8138-c419c2fead38" (UID: "44b3dee8-179f-486a-8138-c419c2fead38"). InnerVolumeSpecName "kube-api-access-7kspm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.403016 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44b3dee8-179f-486a-8138-c419c2fead38-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44b3dee8-179f-486a-8138-c419c2fead38" (UID: "44b3dee8-179f-486a-8138-c419c2fead38"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.407836 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44b3dee8-179f-486a-8138-c419c2fead38-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.408098 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44b3dee8-179f-486a-8138-c419c2fead38-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.408268 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kspm\" (UniqueName: \"kubernetes.io/projected/44b3dee8-179f-486a-8138-c419c2fead38-kube-api-access-7kspm\") on node \"crc\" DevicePath \"\"" Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.509622 4809 generic.go:334] "Generic (PLEG): container finished" podID="44b3dee8-179f-486a-8138-c419c2fead38" containerID="1d1193c07b790cfae0ec63b3f92d103869a4bf9bdf91cd2c39c48405a591a709" exitCode=0 Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.509666 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlbb6" event={"ID":"44b3dee8-179f-486a-8138-c419c2fead38","Type":"ContainerDied","Data":"1d1193c07b790cfae0ec63b3f92d103869a4bf9bdf91cd2c39c48405a591a709"} Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.509694 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlbb6" event={"ID":"44b3dee8-179f-486a-8138-c419c2fead38","Type":"ContainerDied","Data":"f97967c7cdc70bf2765aa1704e4418e6207e6f0e262997035c455bf65227b5fb"} Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.509713 4809 scope.go:117] "RemoveContainer" containerID="1d1193c07b790cfae0ec63b3f92d103869a4bf9bdf91cd2c39c48405a591a709" Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.509768 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xlbb6" Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.530714 4809 scope.go:117] "RemoveContainer" containerID="a10396bae4fd7fde9cf7f2e97a0ed26ac6063d497cf5b109c83e985ae835c21e" Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.549254 4809 scope.go:117] "RemoveContainer" containerID="2d78d16c996ad2a71d6ae647bd53162d9354be184a52bafab845db02fc7003f4" Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.553167 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xlbb6"] Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.560225 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xlbb6"] Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.596769 4809 scope.go:117] "RemoveContainer" containerID="1d1193c07b790cfae0ec63b3f92d103869a4bf9bdf91cd2c39c48405a591a709" Nov 24 07:50:31 crc kubenswrapper[4809]: E1124 07:50:31.597518 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d1193c07b790cfae0ec63b3f92d103869a4bf9bdf91cd2c39c48405a591a709\": container with ID starting with 1d1193c07b790cfae0ec63b3f92d103869a4bf9bdf91cd2c39c48405a591a709 not found: ID does not exist" containerID="1d1193c07b790cfae0ec63b3f92d103869a4bf9bdf91cd2c39c48405a591a709" Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.597556 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d1193c07b790cfae0ec63b3f92d103869a4bf9bdf91cd2c39c48405a591a709"} err="failed to get container status \"1d1193c07b790cfae0ec63b3f92d103869a4bf9bdf91cd2c39c48405a591a709\": rpc error: code = NotFound desc = could not find container \"1d1193c07b790cfae0ec63b3f92d103869a4bf9bdf91cd2c39c48405a591a709\": container with ID starting with 1d1193c07b790cfae0ec63b3f92d103869a4bf9bdf91cd2c39c48405a591a709 not found: ID does not exist" Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.597583 4809 scope.go:117] "RemoveContainer" containerID="a10396bae4fd7fde9cf7f2e97a0ed26ac6063d497cf5b109c83e985ae835c21e" Nov 24 07:50:31 crc kubenswrapper[4809]: E1124 07:50:31.598195 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a10396bae4fd7fde9cf7f2e97a0ed26ac6063d497cf5b109c83e985ae835c21e\": container with ID starting with a10396bae4fd7fde9cf7f2e97a0ed26ac6063d497cf5b109c83e985ae835c21e not found: ID does not exist" containerID="a10396bae4fd7fde9cf7f2e97a0ed26ac6063d497cf5b109c83e985ae835c21e" Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.598217 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a10396bae4fd7fde9cf7f2e97a0ed26ac6063d497cf5b109c83e985ae835c21e"} err="failed to get container status \"a10396bae4fd7fde9cf7f2e97a0ed26ac6063d497cf5b109c83e985ae835c21e\": rpc error: code = NotFound desc = could not find container \"a10396bae4fd7fde9cf7f2e97a0ed26ac6063d497cf5b109c83e985ae835c21e\": container with ID starting with a10396bae4fd7fde9cf7f2e97a0ed26ac6063d497cf5b109c83e985ae835c21e not found: ID does not exist" Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.598230 4809 scope.go:117] "RemoveContainer" containerID="2d78d16c996ad2a71d6ae647bd53162d9354be184a52bafab845db02fc7003f4" Nov 24 07:50:31 crc kubenswrapper[4809]: E1124 07:50:31.598475 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d78d16c996ad2a71d6ae647bd53162d9354be184a52bafab845db02fc7003f4\": container with ID starting with 2d78d16c996ad2a71d6ae647bd53162d9354be184a52bafab845db02fc7003f4 not found: ID does not exist" containerID="2d78d16c996ad2a71d6ae647bd53162d9354be184a52bafab845db02fc7003f4" Nov 24 07:50:31 crc kubenswrapper[4809]: I1124 07:50:31.598500 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d78d16c996ad2a71d6ae647bd53162d9354be184a52bafab845db02fc7003f4"} err="failed to get container status \"2d78d16c996ad2a71d6ae647bd53162d9354be184a52bafab845db02fc7003f4\": rpc error: code = NotFound desc = could not find container \"2d78d16c996ad2a71d6ae647bd53162d9354be184a52bafab845db02fc7003f4\": container with ID starting with 2d78d16c996ad2a71d6ae647bd53162d9354be184a52bafab845db02fc7003f4 not found: ID does not exist" Nov 24 07:50:32 crc kubenswrapper[4809]: I1124 07:50:32.906806 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44b3dee8-179f-486a-8138-c419c2fead38" path="/var/lib/kubelet/pods/44b3dee8-179f-486a-8138-c419c2fead38/volumes" Nov 24 07:50:35 crc kubenswrapper[4809]: I1124 07:50:35.544892 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"19e09f12-9159-4150-ae5b-6d718e6f9957","Type":"ContainerStarted","Data":"daabf5c96fc370cbea539a2f455496667a79060e97b21c74e12fbe952e87624d"} Nov 24 07:50:35 crc kubenswrapper[4809]: I1124 07:50:35.564817 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.6179083680000002 podStartE2EDuration="6.564801001s" podCreationTimestamp="2025-11-24 07:50:29 +0000 UTC" firstStartedPulling="2025-11-24 07:50:30.208029525 +0000 UTC m=+3330.108621230" lastFinishedPulling="2025-11-24 07:50:35.154922128 +0000 UTC m=+3335.055513863" observedRunningTime="2025-11-24 07:50:35.561818241 +0000 UTC m=+3335.462409946" watchObservedRunningTime="2025-11-24 07:50:35.564801001 +0000 UTC m=+3335.465392706" Nov 24 07:50:48 crc kubenswrapper[4809]: I1124 07:50:48.044243 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:50:48 crc kubenswrapper[4809]: I1124 07:50:48.044768 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:50:59 crc kubenswrapper[4809]: I1124 07:50:59.888917 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2bscz/must-gather-d5cpr"] Nov 24 07:50:59 crc kubenswrapper[4809]: E1124 07:50:59.891511 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44b3dee8-179f-486a-8138-c419c2fead38" containerName="extract-utilities" Nov 24 07:50:59 crc kubenswrapper[4809]: I1124 07:50:59.891544 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="44b3dee8-179f-486a-8138-c419c2fead38" containerName="extract-utilities" Nov 24 07:50:59 crc kubenswrapper[4809]: E1124 07:50:59.891564 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44b3dee8-179f-486a-8138-c419c2fead38" containerName="extract-content" Nov 24 07:50:59 crc kubenswrapper[4809]: I1124 07:50:59.891571 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="44b3dee8-179f-486a-8138-c419c2fead38" containerName="extract-content" Nov 24 07:50:59 crc kubenswrapper[4809]: E1124 07:50:59.891608 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44b3dee8-179f-486a-8138-c419c2fead38" containerName="registry-server" Nov 24 07:50:59 crc kubenswrapper[4809]: I1124 07:50:59.891614 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="44b3dee8-179f-486a-8138-c419c2fead38" containerName="registry-server" Nov 24 07:50:59 crc kubenswrapper[4809]: I1124 07:50:59.891874 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="44b3dee8-179f-486a-8138-c419c2fead38" containerName="registry-server" Nov 24 07:50:59 crc kubenswrapper[4809]: I1124 07:50:59.893049 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2bscz/must-gather-d5cpr" Nov 24 07:50:59 crc kubenswrapper[4809]: I1124 07:50:59.896863 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-2bscz"/"openshift-service-ca.crt" Nov 24 07:50:59 crc kubenswrapper[4809]: I1124 07:50:59.903151 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-2bscz"/"kube-root-ca.crt" Nov 24 07:50:59 crc kubenswrapper[4809]: I1124 07:50:59.913145 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-2bscz/must-gather-d5cpr"] Nov 24 07:51:00 crc kubenswrapper[4809]: I1124 07:51:00.019289 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/563bc0df-08a9-42e8-a830-89c568611d6a-must-gather-output\") pod \"must-gather-d5cpr\" (UID: \"563bc0df-08a9-42e8-a830-89c568611d6a\") " pod="openshift-must-gather-2bscz/must-gather-d5cpr" Nov 24 07:51:00 crc kubenswrapper[4809]: I1124 07:51:00.019407 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9clx\" (UniqueName: \"kubernetes.io/projected/563bc0df-08a9-42e8-a830-89c568611d6a-kube-api-access-f9clx\") pod \"must-gather-d5cpr\" (UID: \"563bc0df-08a9-42e8-a830-89c568611d6a\") " pod="openshift-must-gather-2bscz/must-gather-d5cpr" Nov 24 07:51:00 crc kubenswrapper[4809]: I1124 07:51:00.121113 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/563bc0df-08a9-42e8-a830-89c568611d6a-must-gather-output\") pod \"must-gather-d5cpr\" (UID: \"563bc0df-08a9-42e8-a830-89c568611d6a\") " pod="openshift-must-gather-2bscz/must-gather-d5cpr" Nov 24 07:51:00 crc kubenswrapper[4809]: I1124 07:51:00.121472 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9clx\" (UniqueName: \"kubernetes.io/projected/563bc0df-08a9-42e8-a830-89c568611d6a-kube-api-access-f9clx\") pod \"must-gather-d5cpr\" (UID: \"563bc0df-08a9-42e8-a830-89c568611d6a\") " pod="openshift-must-gather-2bscz/must-gather-d5cpr" Nov 24 07:51:00 crc kubenswrapper[4809]: I1124 07:51:00.121804 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/563bc0df-08a9-42e8-a830-89c568611d6a-must-gather-output\") pod \"must-gather-d5cpr\" (UID: \"563bc0df-08a9-42e8-a830-89c568611d6a\") " pod="openshift-must-gather-2bscz/must-gather-d5cpr" Nov 24 07:51:00 crc kubenswrapper[4809]: I1124 07:51:00.143660 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9clx\" (UniqueName: \"kubernetes.io/projected/563bc0df-08a9-42e8-a830-89c568611d6a-kube-api-access-f9clx\") pod \"must-gather-d5cpr\" (UID: \"563bc0df-08a9-42e8-a830-89c568611d6a\") " pod="openshift-must-gather-2bscz/must-gather-d5cpr" Nov 24 07:51:00 crc kubenswrapper[4809]: I1124 07:51:00.222733 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2bscz/must-gather-d5cpr" Nov 24 07:51:00 crc kubenswrapper[4809]: I1124 07:51:00.774121 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-2bscz/must-gather-d5cpr"] Nov 24 07:51:00 crc kubenswrapper[4809]: I1124 07:51:00.806722 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2bscz/must-gather-d5cpr" event={"ID":"563bc0df-08a9-42e8-a830-89c568611d6a","Type":"ContainerStarted","Data":"8f925f58553d7082701ae6b3e09f7a2e08b5b00246338f179c8471782fe8a8a7"} Nov 24 07:51:06 crc kubenswrapper[4809]: I1124 07:51:06.866623 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2bscz/must-gather-d5cpr" event={"ID":"563bc0df-08a9-42e8-a830-89c568611d6a","Type":"ContainerStarted","Data":"3464206eae56260f7d90a77cc37ab4243f12fd71622139823ed2e1c237bcde0c"} Nov 24 07:51:07 crc kubenswrapper[4809]: I1124 07:51:07.877534 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2bscz/must-gather-d5cpr" event={"ID":"563bc0df-08a9-42e8-a830-89c568611d6a","Type":"ContainerStarted","Data":"efc18cc66fb1ea553b0d7463960298ad7dbf6941c29f3160f1bf34af34e8a12c"} Nov 24 07:51:07 crc kubenswrapper[4809]: I1124 07:51:07.897649 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-2bscz/must-gather-d5cpr" podStartSLOduration=3.081775024 podStartE2EDuration="8.897632388s" podCreationTimestamp="2025-11-24 07:50:59 +0000 UTC" firstStartedPulling="2025-11-24 07:51:00.775166942 +0000 UTC m=+3360.675758647" lastFinishedPulling="2025-11-24 07:51:06.591024296 +0000 UTC m=+3366.491616011" observedRunningTime="2025-11-24 07:51:07.892662114 +0000 UTC m=+3367.793253819" watchObservedRunningTime="2025-11-24 07:51:07.897632388 +0000 UTC m=+3367.798224093" Nov 24 07:51:10 crc kubenswrapper[4809]: I1124 07:51:10.210676 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2bscz/crc-debug-rp2fm"] Nov 24 07:51:10 crc kubenswrapper[4809]: I1124 07:51:10.212114 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2bscz/crc-debug-rp2fm" Nov 24 07:51:10 crc kubenswrapper[4809]: I1124 07:51:10.213757 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-2bscz"/"default-dockercfg-r2zpn" Nov 24 07:51:10 crc kubenswrapper[4809]: I1124 07:51:10.342801 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/608e7e93-9675-4e8e-b87b-e1448176f3f9-host\") pod \"crc-debug-rp2fm\" (UID: \"608e7e93-9675-4e8e-b87b-e1448176f3f9\") " pod="openshift-must-gather-2bscz/crc-debug-rp2fm" Nov 24 07:51:10 crc kubenswrapper[4809]: I1124 07:51:10.343245 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77j45\" (UniqueName: \"kubernetes.io/projected/608e7e93-9675-4e8e-b87b-e1448176f3f9-kube-api-access-77j45\") pod \"crc-debug-rp2fm\" (UID: \"608e7e93-9675-4e8e-b87b-e1448176f3f9\") " pod="openshift-must-gather-2bscz/crc-debug-rp2fm" Nov 24 07:51:10 crc kubenswrapper[4809]: I1124 07:51:10.445100 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77j45\" (UniqueName: \"kubernetes.io/projected/608e7e93-9675-4e8e-b87b-e1448176f3f9-kube-api-access-77j45\") pod \"crc-debug-rp2fm\" (UID: \"608e7e93-9675-4e8e-b87b-e1448176f3f9\") " pod="openshift-must-gather-2bscz/crc-debug-rp2fm" Nov 24 07:51:10 crc kubenswrapper[4809]: I1124 07:51:10.445186 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/608e7e93-9675-4e8e-b87b-e1448176f3f9-host\") pod \"crc-debug-rp2fm\" (UID: \"608e7e93-9675-4e8e-b87b-e1448176f3f9\") " pod="openshift-must-gather-2bscz/crc-debug-rp2fm" Nov 24 07:51:10 crc kubenswrapper[4809]: I1124 07:51:10.445266 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/608e7e93-9675-4e8e-b87b-e1448176f3f9-host\") pod \"crc-debug-rp2fm\" (UID: \"608e7e93-9675-4e8e-b87b-e1448176f3f9\") " pod="openshift-must-gather-2bscz/crc-debug-rp2fm" Nov 24 07:51:10 crc kubenswrapper[4809]: I1124 07:51:10.479769 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77j45\" (UniqueName: \"kubernetes.io/projected/608e7e93-9675-4e8e-b87b-e1448176f3f9-kube-api-access-77j45\") pod \"crc-debug-rp2fm\" (UID: \"608e7e93-9675-4e8e-b87b-e1448176f3f9\") " pod="openshift-must-gather-2bscz/crc-debug-rp2fm" Nov 24 07:51:10 crc kubenswrapper[4809]: I1124 07:51:10.533122 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2bscz/crc-debug-rp2fm" Nov 24 07:51:10 crc kubenswrapper[4809]: I1124 07:51:10.905693 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2bscz/crc-debug-rp2fm" event={"ID":"608e7e93-9675-4e8e-b87b-e1448176f3f9","Type":"ContainerStarted","Data":"cd5777e03a8fc1fea36d000259c3785ac9eadfef7e6057349b497b068f3da881"} Nov 24 07:51:18 crc kubenswrapper[4809]: I1124 07:51:18.043226 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:51:18 crc kubenswrapper[4809]: I1124 07:51:18.043807 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:51:22 crc kubenswrapper[4809]: I1124 07:51:22.011775 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2bscz/crc-debug-rp2fm" event={"ID":"608e7e93-9675-4e8e-b87b-e1448176f3f9","Type":"ContainerStarted","Data":"1e572b1c76177f14c9977fc4104ab2a9b1198238fac01678e94232d76bd74005"} Nov 24 07:51:38 crc kubenswrapper[4809]: I1124 07:51:38.603269 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-2bscz/crc-debug-rp2fm" podStartSLOduration=17.762986653 podStartE2EDuration="28.603252545s" podCreationTimestamp="2025-11-24 07:51:10 +0000 UTC" firstStartedPulling="2025-11-24 07:51:10.569170112 +0000 UTC m=+3370.469761817" lastFinishedPulling="2025-11-24 07:51:21.409435994 +0000 UTC m=+3381.310027709" observedRunningTime="2025-11-24 07:51:22.032441775 +0000 UTC m=+3381.933033480" watchObservedRunningTime="2025-11-24 07:51:38.603252545 +0000 UTC m=+3398.503844250" Nov 24 07:51:38 crc kubenswrapper[4809]: I1124 07:51:38.611080 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qrkrx"] Nov 24 07:51:38 crc kubenswrapper[4809]: I1124 07:51:38.613293 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qrkrx" Nov 24 07:51:38 crc kubenswrapper[4809]: I1124 07:51:38.624604 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qrkrx"] Nov 24 07:51:38 crc kubenswrapper[4809]: I1124 07:51:38.789391 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwlpl\" (UniqueName: \"kubernetes.io/projected/38722d08-fe98-4aaa-9a62-7a7c3e34d024-kube-api-access-dwlpl\") pod \"certified-operators-qrkrx\" (UID: \"38722d08-fe98-4aaa-9a62-7a7c3e34d024\") " pod="openshift-marketplace/certified-operators-qrkrx" Nov 24 07:51:38 crc kubenswrapper[4809]: I1124 07:51:38.789533 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38722d08-fe98-4aaa-9a62-7a7c3e34d024-catalog-content\") pod \"certified-operators-qrkrx\" (UID: \"38722d08-fe98-4aaa-9a62-7a7c3e34d024\") " pod="openshift-marketplace/certified-operators-qrkrx" Nov 24 07:51:38 crc kubenswrapper[4809]: I1124 07:51:38.789692 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38722d08-fe98-4aaa-9a62-7a7c3e34d024-utilities\") pod \"certified-operators-qrkrx\" (UID: \"38722d08-fe98-4aaa-9a62-7a7c3e34d024\") " pod="openshift-marketplace/certified-operators-qrkrx" Nov 24 07:51:38 crc kubenswrapper[4809]: I1124 07:51:38.892342 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwlpl\" (UniqueName: \"kubernetes.io/projected/38722d08-fe98-4aaa-9a62-7a7c3e34d024-kube-api-access-dwlpl\") pod \"certified-operators-qrkrx\" (UID: \"38722d08-fe98-4aaa-9a62-7a7c3e34d024\") " pod="openshift-marketplace/certified-operators-qrkrx" Nov 24 07:51:38 crc kubenswrapper[4809]: I1124 07:51:38.892727 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38722d08-fe98-4aaa-9a62-7a7c3e34d024-catalog-content\") pod \"certified-operators-qrkrx\" (UID: \"38722d08-fe98-4aaa-9a62-7a7c3e34d024\") " pod="openshift-marketplace/certified-operators-qrkrx" Nov 24 07:51:38 crc kubenswrapper[4809]: I1124 07:51:38.892771 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38722d08-fe98-4aaa-9a62-7a7c3e34d024-utilities\") pod \"certified-operators-qrkrx\" (UID: \"38722d08-fe98-4aaa-9a62-7a7c3e34d024\") " pod="openshift-marketplace/certified-operators-qrkrx" Nov 24 07:51:38 crc kubenswrapper[4809]: I1124 07:51:38.893245 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38722d08-fe98-4aaa-9a62-7a7c3e34d024-catalog-content\") pod \"certified-operators-qrkrx\" (UID: \"38722d08-fe98-4aaa-9a62-7a7c3e34d024\") " pod="openshift-marketplace/certified-operators-qrkrx" Nov 24 07:51:38 crc kubenswrapper[4809]: I1124 07:51:38.893247 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38722d08-fe98-4aaa-9a62-7a7c3e34d024-utilities\") pod \"certified-operators-qrkrx\" (UID: \"38722d08-fe98-4aaa-9a62-7a7c3e34d024\") " pod="openshift-marketplace/certified-operators-qrkrx" Nov 24 07:51:38 crc kubenswrapper[4809]: I1124 07:51:38.915872 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwlpl\" (UniqueName: \"kubernetes.io/projected/38722d08-fe98-4aaa-9a62-7a7c3e34d024-kube-api-access-dwlpl\") pod \"certified-operators-qrkrx\" (UID: \"38722d08-fe98-4aaa-9a62-7a7c3e34d024\") " pod="openshift-marketplace/certified-operators-qrkrx" Nov 24 07:51:38 crc kubenswrapper[4809]: I1124 07:51:38.934424 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qrkrx" Nov 24 07:51:40 crc kubenswrapper[4809]: I1124 07:51:40.193899 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qrkrx"] Nov 24 07:51:41 crc kubenswrapper[4809]: I1124 07:51:41.188061 4809 generic.go:334] "Generic (PLEG): container finished" podID="38722d08-fe98-4aaa-9a62-7a7c3e34d024" containerID="926f6a8850d771de4908601d30d99fb7ca55885ff91c78811afd526df694db5c" exitCode=0 Nov 24 07:51:41 crc kubenswrapper[4809]: I1124 07:51:41.188244 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qrkrx" event={"ID":"38722d08-fe98-4aaa-9a62-7a7c3e34d024","Type":"ContainerDied","Data":"926f6a8850d771de4908601d30d99fb7ca55885ff91c78811afd526df694db5c"} Nov 24 07:51:41 crc kubenswrapper[4809]: I1124 07:51:41.188690 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qrkrx" event={"ID":"38722d08-fe98-4aaa-9a62-7a7c3e34d024","Type":"ContainerStarted","Data":"efbc9e14345251f8a598a7271aa297e5f8b4ad4d76af795a111f30f200cea366"} Nov 24 07:51:43 crc kubenswrapper[4809]: I1124 07:51:43.207519 4809 generic.go:334] "Generic (PLEG): container finished" podID="38722d08-fe98-4aaa-9a62-7a7c3e34d024" containerID="ede2c52421db16a1bbaae8af84c1a3e777004d53324712c6107eb039855b93b9" exitCode=0 Nov 24 07:51:43 crc kubenswrapper[4809]: I1124 07:51:43.208088 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qrkrx" event={"ID":"38722d08-fe98-4aaa-9a62-7a7c3e34d024","Type":"ContainerDied","Data":"ede2c52421db16a1bbaae8af84c1a3e777004d53324712c6107eb039855b93b9"} Nov 24 07:51:44 crc kubenswrapper[4809]: I1124 07:51:44.218471 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qrkrx" event={"ID":"38722d08-fe98-4aaa-9a62-7a7c3e34d024","Type":"ContainerStarted","Data":"63578127fa33ddf040f135935dd741769c7b8bd239ee0e3cd843edc475ea1c96"} Nov 24 07:51:44 crc kubenswrapper[4809]: I1124 07:51:44.237889 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qrkrx" podStartSLOduration=3.4259928889999998 podStartE2EDuration="6.23787014s" podCreationTimestamp="2025-11-24 07:51:38 +0000 UTC" firstStartedPulling="2025-11-24 07:51:41.1912277 +0000 UTC m=+3401.091819415" lastFinishedPulling="2025-11-24 07:51:44.003104961 +0000 UTC m=+3403.903696666" observedRunningTime="2025-11-24 07:51:44.235129146 +0000 UTC m=+3404.135720861" watchObservedRunningTime="2025-11-24 07:51:44.23787014 +0000 UTC m=+3404.138461845" Nov 24 07:51:48 crc kubenswrapper[4809]: I1124 07:51:48.042801 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:51:48 crc kubenswrapper[4809]: I1124 07:51:48.043122 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:51:48 crc kubenswrapper[4809]: I1124 07:51:48.043165 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 07:51:48 crc kubenswrapper[4809]: I1124 07:51:48.043762 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7d5c0ef1309a4cda8c3a374bbff14973c66d06a10b1ee334b7a60d88f824f490"} pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:51:48 crc kubenswrapper[4809]: I1124 07:51:48.043804 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" containerID="cri-o://7d5c0ef1309a4cda8c3a374bbff14973c66d06a10b1ee334b7a60d88f824f490" gracePeriod=600 Nov 24 07:51:48 crc kubenswrapper[4809]: I1124 07:51:48.256587 4809 generic.go:334] "Generic (PLEG): container finished" podID="be8050d5-1aea-453e-a06e-87702b0e856b" containerID="7d5c0ef1309a4cda8c3a374bbff14973c66d06a10b1ee334b7a60d88f824f490" exitCode=0 Nov 24 07:51:48 crc kubenswrapper[4809]: I1124 07:51:48.256636 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerDied","Data":"7d5c0ef1309a4cda8c3a374bbff14973c66d06a10b1ee334b7a60d88f824f490"} Nov 24 07:51:48 crc kubenswrapper[4809]: I1124 07:51:48.256669 4809 scope.go:117] "RemoveContainer" containerID="1c07db345cef3d165b80880e80550d46baa252264ed80cff2aad46181090e759" Nov 24 07:51:48 crc kubenswrapper[4809]: I1124 07:51:48.935355 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qrkrx" Nov 24 07:51:48 crc kubenswrapper[4809]: I1124 07:51:48.936008 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qrkrx" Nov 24 07:51:48 crc kubenswrapper[4809]: I1124 07:51:48.992423 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qrkrx" Nov 24 07:51:49 crc kubenswrapper[4809]: I1124 07:51:49.266869 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerStarted","Data":"2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0"} Nov 24 07:51:49 crc kubenswrapper[4809]: I1124 07:51:49.315896 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qrkrx" Nov 24 07:51:49 crc kubenswrapper[4809]: I1124 07:51:49.363346 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qrkrx"] Nov 24 07:51:51 crc kubenswrapper[4809]: I1124 07:51:51.288061 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qrkrx" podUID="38722d08-fe98-4aaa-9a62-7a7c3e34d024" containerName="registry-server" containerID="cri-o://63578127fa33ddf040f135935dd741769c7b8bd239ee0e3cd843edc475ea1c96" gracePeriod=2 Nov 24 07:51:51 crc kubenswrapper[4809]: I1124 07:51:51.711533 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qrkrx" Nov 24 07:51:51 crc kubenswrapper[4809]: I1124 07:51:51.833384 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38722d08-fe98-4aaa-9a62-7a7c3e34d024-utilities\") pod \"38722d08-fe98-4aaa-9a62-7a7c3e34d024\" (UID: \"38722d08-fe98-4aaa-9a62-7a7c3e34d024\") " Nov 24 07:51:51 crc kubenswrapper[4809]: I1124 07:51:51.833538 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwlpl\" (UniqueName: \"kubernetes.io/projected/38722d08-fe98-4aaa-9a62-7a7c3e34d024-kube-api-access-dwlpl\") pod \"38722d08-fe98-4aaa-9a62-7a7c3e34d024\" (UID: \"38722d08-fe98-4aaa-9a62-7a7c3e34d024\") " Nov 24 07:51:51 crc kubenswrapper[4809]: I1124 07:51:51.833592 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38722d08-fe98-4aaa-9a62-7a7c3e34d024-catalog-content\") pod \"38722d08-fe98-4aaa-9a62-7a7c3e34d024\" (UID: \"38722d08-fe98-4aaa-9a62-7a7c3e34d024\") " Nov 24 07:51:51 crc kubenswrapper[4809]: I1124 07:51:51.834437 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38722d08-fe98-4aaa-9a62-7a7c3e34d024-utilities" (OuterVolumeSpecName: "utilities") pod "38722d08-fe98-4aaa-9a62-7a7c3e34d024" (UID: "38722d08-fe98-4aaa-9a62-7a7c3e34d024"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:51:51 crc kubenswrapper[4809]: I1124 07:51:51.838837 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38722d08-fe98-4aaa-9a62-7a7c3e34d024-kube-api-access-dwlpl" (OuterVolumeSpecName: "kube-api-access-dwlpl") pod "38722d08-fe98-4aaa-9a62-7a7c3e34d024" (UID: "38722d08-fe98-4aaa-9a62-7a7c3e34d024"). InnerVolumeSpecName "kube-api-access-dwlpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:51:51 crc kubenswrapper[4809]: I1124 07:51:51.880844 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38722d08-fe98-4aaa-9a62-7a7c3e34d024-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38722d08-fe98-4aaa-9a62-7a7c3e34d024" (UID: "38722d08-fe98-4aaa-9a62-7a7c3e34d024"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:51:51 crc kubenswrapper[4809]: I1124 07:51:51.935309 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38722d08-fe98-4aaa-9a62-7a7c3e34d024-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:51:51 crc kubenswrapper[4809]: I1124 07:51:51.935343 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwlpl\" (UniqueName: \"kubernetes.io/projected/38722d08-fe98-4aaa-9a62-7a7c3e34d024-kube-api-access-dwlpl\") on node \"crc\" DevicePath \"\"" Nov 24 07:51:51 crc kubenswrapper[4809]: I1124 07:51:51.935353 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38722d08-fe98-4aaa-9a62-7a7c3e34d024-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:51:52 crc kubenswrapper[4809]: I1124 07:51:52.300875 4809 generic.go:334] "Generic (PLEG): container finished" podID="38722d08-fe98-4aaa-9a62-7a7c3e34d024" containerID="63578127fa33ddf040f135935dd741769c7b8bd239ee0e3cd843edc475ea1c96" exitCode=0 Nov 24 07:51:52 crc kubenswrapper[4809]: I1124 07:51:52.300978 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qrkrx" event={"ID":"38722d08-fe98-4aaa-9a62-7a7c3e34d024","Type":"ContainerDied","Data":"63578127fa33ddf040f135935dd741769c7b8bd239ee0e3cd843edc475ea1c96"} Nov 24 07:51:52 crc kubenswrapper[4809]: I1124 07:51:52.301206 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qrkrx" event={"ID":"38722d08-fe98-4aaa-9a62-7a7c3e34d024","Type":"ContainerDied","Data":"efbc9e14345251f8a598a7271aa297e5f8b4ad4d76af795a111f30f200cea366"} Nov 24 07:51:52 crc kubenswrapper[4809]: I1124 07:51:52.301233 4809 scope.go:117] "RemoveContainer" containerID="63578127fa33ddf040f135935dd741769c7b8bd239ee0e3cd843edc475ea1c96" Nov 24 07:51:52 crc kubenswrapper[4809]: I1124 07:51:52.301006 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qrkrx" Nov 24 07:51:52 crc kubenswrapper[4809]: I1124 07:51:52.322915 4809 scope.go:117] "RemoveContainer" containerID="ede2c52421db16a1bbaae8af84c1a3e777004d53324712c6107eb039855b93b9" Nov 24 07:51:52 crc kubenswrapper[4809]: I1124 07:51:52.336806 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qrkrx"] Nov 24 07:51:52 crc kubenswrapper[4809]: I1124 07:51:52.352889 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qrkrx"] Nov 24 07:51:52 crc kubenswrapper[4809]: I1124 07:51:52.354916 4809 scope.go:117] "RemoveContainer" containerID="926f6a8850d771de4908601d30d99fb7ca55885ff91c78811afd526df694db5c" Nov 24 07:51:52 crc kubenswrapper[4809]: I1124 07:51:52.390585 4809 scope.go:117] "RemoveContainer" containerID="63578127fa33ddf040f135935dd741769c7b8bd239ee0e3cd843edc475ea1c96" Nov 24 07:51:52 crc kubenswrapper[4809]: E1124 07:51:52.390957 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63578127fa33ddf040f135935dd741769c7b8bd239ee0e3cd843edc475ea1c96\": container with ID starting with 63578127fa33ddf040f135935dd741769c7b8bd239ee0e3cd843edc475ea1c96 not found: ID does not exist" containerID="63578127fa33ddf040f135935dd741769c7b8bd239ee0e3cd843edc475ea1c96" Nov 24 07:51:52 crc kubenswrapper[4809]: I1124 07:51:52.391011 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63578127fa33ddf040f135935dd741769c7b8bd239ee0e3cd843edc475ea1c96"} err="failed to get container status \"63578127fa33ddf040f135935dd741769c7b8bd239ee0e3cd843edc475ea1c96\": rpc error: code = NotFound desc = could not find container \"63578127fa33ddf040f135935dd741769c7b8bd239ee0e3cd843edc475ea1c96\": container with ID starting with 63578127fa33ddf040f135935dd741769c7b8bd239ee0e3cd843edc475ea1c96 not found: ID does not exist" Nov 24 07:51:52 crc kubenswrapper[4809]: I1124 07:51:52.391039 4809 scope.go:117] "RemoveContainer" containerID="ede2c52421db16a1bbaae8af84c1a3e777004d53324712c6107eb039855b93b9" Nov 24 07:51:52 crc kubenswrapper[4809]: E1124 07:51:52.391378 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ede2c52421db16a1bbaae8af84c1a3e777004d53324712c6107eb039855b93b9\": container with ID starting with ede2c52421db16a1bbaae8af84c1a3e777004d53324712c6107eb039855b93b9 not found: ID does not exist" containerID="ede2c52421db16a1bbaae8af84c1a3e777004d53324712c6107eb039855b93b9" Nov 24 07:51:52 crc kubenswrapper[4809]: I1124 07:51:52.391420 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ede2c52421db16a1bbaae8af84c1a3e777004d53324712c6107eb039855b93b9"} err="failed to get container status \"ede2c52421db16a1bbaae8af84c1a3e777004d53324712c6107eb039855b93b9\": rpc error: code = NotFound desc = could not find container \"ede2c52421db16a1bbaae8af84c1a3e777004d53324712c6107eb039855b93b9\": container with ID starting with ede2c52421db16a1bbaae8af84c1a3e777004d53324712c6107eb039855b93b9 not found: ID does not exist" Nov 24 07:51:52 crc kubenswrapper[4809]: I1124 07:51:52.391456 4809 scope.go:117] "RemoveContainer" containerID="926f6a8850d771de4908601d30d99fb7ca55885ff91c78811afd526df694db5c" Nov 24 07:51:52 crc kubenswrapper[4809]: E1124 07:51:52.391721 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"926f6a8850d771de4908601d30d99fb7ca55885ff91c78811afd526df694db5c\": container with ID starting with 926f6a8850d771de4908601d30d99fb7ca55885ff91c78811afd526df694db5c not found: ID does not exist" containerID="926f6a8850d771de4908601d30d99fb7ca55885ff91c78811afd526df694db5c" Nov 24 07:51:52 crc kubenswrapper[4809]: I1124 07:51:52.391743 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"926f6a8850d771de4908601d30d99fb7ca55885ff91c78811afd526df694db5c"} err="failed to get container status \"926f6a8850d771de4908601d30d99fb7ca55885ff91c78811afd526df694db5c\": rpc error: code = NotFound desc = could not find container \"926f6a8850d771de4908601d30d99fb7ca55885ff91c78811afd526df694db5c\": container with ID starting with 926f6a8850d771de4908601d30d99fb7ca55885ff91c78811afd526df694db5c not found: ID does not exist" Nov 24 07:51:52 crc kubenswrapper[4809]: I1124 07:51:52.903262 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38722d08-fe98-4aaa-9a62-7a7c3e34d024" path="/var/lib/kubelet/pods/38722d08-fe98-4aaa-9a62-7a7c3e34d024/volumes" Nov 24 07:52:02 crc kubenswrapper[4809]: I1124 07:52:02.421530 4809 generic.go:334] "Generic (PLEG): container finished" podID="608e7e93-9675-4e8e-b87b-e1448176f3f9" containerID="1e572b1c76177f14c9977fc4104ab2a9b1198238fac01678e94232d76bd74005" exitCode=0 Nov 24 07:52:02 crc kubenswrapper[4809]: I1124 07:52:02.421636 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2bscz/crc-debug-rp2fm" event={"ID":"608e7e93-9675-4e8e-b87b-e1448176f3f9","Type":"ContainerDied","Data":"1e572b1c76177f14c9977fc4104ab2a9b1198238fac01678e94232d76bd74005"} Nov 24 07:52:03 crc kubenswrapper[4809]: I1124 07:52:03.549536 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2bscz/crc-debug-rp2fm" Nov 24 07:52:03 crc kubenswrapper[4809]: I1124 07:52:03.588440 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2bscz/crc-debug-rp2fm"] Nov 24 07:52:03 crc kubenswrapper[4809]: I1124 07:52:03.595078 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2bscz/crc-debug-rp2fm"] Nov 24 07:52:03 crc kubenswrapper[4809]: I1124 07:52:03.652193 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77j45\" (UniqueName: \"kubernetes.io/projected/608e7e93-9675-4e8e-b87b-e1448176f3f9-kube-api-access-77j45\") pod \"608e7e93-9675-4e8e-b87b-e1448176f3f9\" (UID: \"608e7e93-9675-4e8e-b87b-e1448176f3f9\") " Nov 24 07:52:03 crc kubenswrapper[4809]: I1124 07:52:03.652481 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/608e7e93-9675-4e8e-b87b-e1448176f3f9-host\") pod \"608e7e93-9675-4e8e-b87b-e1448176f3f9\" (UID: \"608e7e93-9675-4e8e-b87b-e1448176f3f9\") " Nov 24 07:52:03 crc kubenswrapper[4809]: I1124 07:52:03.652673 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/608e7e93-9675-4e8e-b87b-e1448176f3f9-host" (OuterVolumeSpecName: "host") pod "608e7e93-9675-4e8e-b87b-e1448176f3f9" (UID: "608e7e93-9675-4e8e-b87b-e1448176f3f9"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:52:03 crc kubenswrapper[4809]: I1124 07:52:03.653459 4809 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/608e7e93-9675-4e8e-b87b-e1448176f3f9-host\") on node \"crc\" DevicePath \"\"" Nov 24 07:52:03 crc kubenswrapper[4809]: I1124 07:52:03.660188 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/608e7e93-9675-4e8e-b87b-e1448176f3f9-kube-api-access-77j45" (OuterVolumeSpecName: "kube-api-access-77j45") pod "608e7e93-9675-4e8e-b87b-e1448176f3f9" (UID: "608e7e93-9675-4e8e-b87b-e1448176f3f9"). InnerVolumeSpecName "kube-api-access-77j45". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:52:03 crc kubenswrapper[4809]: I1124 07:52:03.754757 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77j45\" (UniqueName: \"kubernetes.io/projected/608e7e93-9675-4e8e-b87b-e1448176f3f9-kube-api-access-77j45\") on node \"crc\" DevicePath \"\"" Nov 24 07:52:04 crc kubenswrapper[4809]: I1124 07:52:04.444948 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd5777e03a8fc1fea36d000259c3785ac9eadfef7e6057349b497b068f3da881" Nov 24 07:52:04 crc kubenswrapper[4809]: I1124 07:52:04.445472 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2bscz/crc-debug-rp2fm" Nov 24 07:52:04 crc kubenswrapper[4809]: I1124 07:52:04.826500 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2bscz/crc-debug-6rtvd"] Nov 24 07:52:04 crc kubenswrapper[4809]: E1124 07:52:04.826924 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="608e7e93-9675-4e8e-b87b-e1448176f3f9" containerName="container-00" Nov 24 07:52:04 crc kubenswrapper[4809]: I1124 07:52:04.826941 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="608e7e93-9675-4e8e-b87b-e1448176f3f9" containerName="container-00" Nov 24 07:52:04 crc kubenswrapper[4809]: E1124 07:52:04.826954 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38722d08-fe98-4aaa-9a62-7a7c3e34d024" containerName="registry-server" Nov 24 07:52:04 crc kubenswrapper[4809]: I1124 07:52:04.826976 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="38722d08-fe98-4aaa-9a62-7a7c3e34d024" containerName="registry-server" Nov 24 07:52:04 crc kubenswrapper[4809]: E1124 07:52:04.826986 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38722d08-fe98-4aaa-9a62-7a7c3e34d024" containerName="extract-utilities" Nov 24 07:52:04 crc kubenswrapper[4809]: I1124 07:52:04.826994 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="38722d08-fe98-4aaa-9a62-7a7c3e34d024" containerName="extract-utilities" Nov 24 07:52:04 crc kubenswrapper[4809]: E1124 07:52:04.827010 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38722d08-fe98-4aaa-9a62-7a7c3e34d024" containerName="extract-content" Nov 24 07:52:04 crc kubenswrapper[4809]: I1124 07:52:04.827016 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="38722d08-fe98-4aaa-9a62-7a7c3e34d024" containerName="extract-content" Nov 24 07:52:04 crc kubenswrapper[4809]: I1124 07:52:04.827220 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="38722d08-fe98-4aaa-9a62-7a7c3e34d024" containerName="registry-server" Nov 24 07:52:04 crc kubenswrapper[4809]: I1124 07:52:04.827245 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="608e7e93-9675-4e8e-b87b-e1448176f3f9" containerName="container-00" Nov 24 07:52:04 crc kubenswrapper[4809]: I1124 07:52:04.827884 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2bscz/crc-debug-6rtvd" Nov 24 07:52:04 crc kubenswrapper[4809]: I1124 07:52:04.830351 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-2bscz"/"default-dockercfg-r2zpn" Nov 24 07:52:04 crc kubenswrapper[4809]: I1124 07:52:04.878485 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8s9p\" (UniqueName: \"kubernetes.io/projected/f940c5cc-383e-4a50-be80-ff715acc87fc-kube-api-access-j8s9p\") pod \"crc-debug-6rtvd\" (UID: \"f940c5cc-383e-4a50-be80-ff715acc87fc\") " pod="openshift-must-gather-2bscz/crc-debug-6rtvd" Nov 24 07:52:04 crc kubenswrapper[4809]: I1124 07:52:04.878691 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f940c5cc-383e-4a50-be80-ff715acc87fc-host\") pod \"crc-debug-6rtvd\" (UID: \"f940c5cc-383e-4a50-be80-ff715acc87fc\") " pod="openshift-must-gather-2bscz/crc-debug-6rtvd" Nov 24 07:52:04 crc kubenswrapper[4809]: I1124 07:52:04.910359 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="608e7e93-9675-4e8e-b87b-e1448176f3f9" path="/var/lib/kubelet/pods/608e7e93-9675-4e8e-b87b-e1448176f3f9/volumes" Nov 24 07:52:04 crc kubenswrapper[4809]: I1124 07:52:04.981212 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8s9p\" (UniqueName: \"kubernetes.io/projected/f940c5cc-383e-4a50-be80-ff715acc87fc-kube-api-access-j8s9p\") pod \"crc-debug-6rtvd\" (UID: \"f940c5cc-383e-4a50-be80-ff715acc87fc\") " pod="openshift-must-gather-2bscz/crc-debug-6rtvd" Nov 24 07:52:04 crc kubenswrapper[4809]: I1124 07:52:04.981325 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f940c5cc-383e-4a50-be80-ff715acc87fc-host\") pod \"crc-debug-6rtvd\" (UID: \"f940c5cc-383e-4a50-be80-ff715acc87fc\") " pod="openshift-must-gather-2bscz/crc-debug-6rtvd" Nov 24 07:52:04 crc kubenswrapper[4809]: I1124 07:52:04.981489 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f940c5cc-383e-4a50-be80-ff715acc87fc-host\") pod \"crc-debug-6rtvd\" (UID: \"f940c5cc-383e-4a50-be80-ff715acc87fc\") " pod="openshift-must-gather-2bscz/crc-debug-6rtvd" Nov 24 07:52:05 crc kubenswrapper[4809]: I1124 07:52:05.003627 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8s9p\" (UniqueName: \"kubernetes.io/projected/f940c5cc-383e-4a50-be80-ff715acc87fc-kube-api-access-j8s9p\") pod \"crc-debug-6rtvd\" (UID: \"f940c5cc-383e-4a50-be80-ff715acc87fc\") " pod="openshift-must-gather-2bscz/crc-debug-6rtvd" Nov 24 07:52:05 crc kubenswrapper[4809]: I1124 07:52:05.151141 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2bscz/crc-debug-6rtvd" Nov 24 07:52:05 crc kubenswrapper[4809]: I1124 07:52:05.456120 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2bscz/crc-debug-6rtvd" event={"ID":"f940c5cc-383e-4a50-be80-ff715acc87fc","Type":"ContainerStarted","Data":"42a03d852bb27ffacc27c7e4314c7008cf983f04875507399491c453d672a257"} Nov 24 07:52:05 crc kubenswrapper[4809]: I1124 07:52:05.456572 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2bscz/crc-debug-6rtvd" event={"ID":"f940c5cc-383e-4a50-be80-ff715acc87fc","Type":"ContainerStarted","Data":"fbe0e72268ff57d4a435b02559f1a66a264b0d20b58f18dc08a3bb9781c807b1"} Nov 24 07:52:05 crc kubenswrapper[4809]: I1124 07:52:05.955650 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2bscz/crc-debug-6rtvd"] Nov 24 07:52:05 crc kubenswrapper[4809]: I1124 07:52:05.962521 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2bscz/crc-debug-6rtvd"] Nov 24 07:52:06 crc kubenswrapper[4809]: I1124 07:52:06.469338 4809 generic.go:334] "Generic (PLEG): container finished" podID="f940c5cc-383e-4a50-be80-ff715acc87fc" containerID="42a03d852bb27ffacc27c7e4314c7008cf983f04875507399491c453d672a257" exitCode=0 Nov 24 07:52:06 crc kubenswrapper[4809]: I1124 07:52:06.576292 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2bscz/crc-debug-6rtvd" Nov 24 07:52:06 crc kubenswrapper[4809]: I1124 07:52:06.718322 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f940c5cc-383e-4a50-be80-ff715acc87fc-host\") pod \"f940c5cc-383e-4a50-be80-ff715acc87fc\" (UID: \"f940c5cc-383e-4a50-be80-ff715acc87fc\") " Nov 24 07:52:06 crc kubenswrapper[4809]: I1124 07:52:06.718451 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f940c5cc-383e-4a50-be80-ff715acc87fc-host" (OuterVolumeSpecName: "host") pod "f940c5cc-383e-4a50-be80-ff715acc87fc" (UID: "f940c5cc-383e-4a50-be80-ff715acc87fc"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:52:06 crc kubenswrapper[4809]: I1124 07:52:06.718612 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8s9p\" (UniqueName: \"kubernetes.io/projected/f940c5cc-383e-4a50-be80-ff715acc87fc-kube-api-access-j8s9p\") pod \"f940c5cc-383e-4a50-be80-ff715acc87fc\" (UID: \"f940c5cc-383e-4a50-be80-ff715acc87fc\") " Nov 24 07:52:06 crc kubenswrapper[4809]: I1124 07:52:06.719141 4809 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f940c5cc-383e-4a50-be80-ff715acc87fc-host\") on node \"crc\" DevicePath \"\"" Nov 24 07:52:06 crc kubenswrapper[4809]: I1124 07:52:06.724179 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f940c5cc-383e-4a50-be80-ff715acc87fc-kube-api-access-j8s9p" (OuterVolumeSpecName: "kube-api-access-j8s9p") pod "f940c5cc-383e-4a50-be80-ff715acc87fc" (UID: "f940c5cc-383e-4a50-be80-ff715acc87fc"). InnerVolumeSpecName "kube-api-access-j8s9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:52:06 crc kubenswrapper[4809]: I1124 07:52:06.820900 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8s9p\" (UniqueName: \"kubernetes.io/projected/f940c5cc-383e-4a50-be80-ff715acc87fc-kube-api-access-j8s9p\") on node \"crc\" DevicePath \"\"" Nov 24 07:52:06 crc kubenswrapper[4809]: I1124 07:52:06.901116 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f940c5cc-383e-4a50-be80-ff715acc87fc" path="/var/lib/kubelet/pods/f940c5cc-383e-4a50-be80-ff715acc87fc/volumes" Nov 24 07:52:07 crc kubenswrapper[4809]: I1124 07:52:07.105162 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2bscz/crc-debug-ccpkj"] Nov 24 07:52:07 crc kubenswrapper[4809]: E1124 07:52:07.105603 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f940c5cc-383e-4a50-be80-ff715acc87fc" containerName="container-00" Nov 24 07:52:07 crc kubenswrapper[4809]: I1124 07:52:07.105616 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f940c5cc-383e-4a50-be80-ff715acc87fc" containerName="container-00" Nov 24 07:52:07 crc kubenswrapper[4809]: I1124 07:52:07.105818 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f940c5cc-383e-4a50-be80-ff715acc87fc" containerName="container-00" Nov 24 07:52:07 crc kubenswrapper[4809]: I1124 07:52:07.106444 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2bscz/crc-debug-ccpkj" Nov 24 07:52:07 crc kubenswrapper[4809]: I1124 07:52:07.227026 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7tkm\" (UniqueName: \"kubernetes.io/projected/fde9c043-8307-4382-8ae1-6fd4114d938d-kube-api-access-f7tkm\") pod \"crc-debug-ccpkj\" (UID: \"fde9c043-8307-4382-8ae1-6fd4114d938d\") " pod="openshift-must-gather-2bscz/crc-debug-ccpkj" Nov 24 07:52:07 crc kubenswrapper[4809]: I1124 07:52:07.227068 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fde9c043-8307-4382-8ae1-6fd4114d938d-host\") pod \"crc-debug-ccpkj\" (UID: \"fde9c043-8307-4382-8ae1-6fd4114d938d\") " pod="openshift-must-gather-2bscz/crc-debug-ccpkj" Nov 24 07:52:07 crc kubenswrapper[4809]: I1124 07:52:07.328517 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7tkm\" (UniqueName: \"kubernetes.io/projected/fde9c043-8307-4382-8ae1-6fd4114d938d-kube-api-access-f7tkm\") pod \"crc-debug-ccpkj\" (UID: \"fde9c043-8307-4382-8ae1-6fd4114d938d\") " pod="openshift-must-gather-2bscz/crc-debug-ccpkj" Nov 24 07:52:07 crc kubenswrapper[4809]: I1124 07:52:07.328563 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fde9c043-8307-4382-8ae1-6fd4114d938d-host\") pod \"crc-debug-ccpkj\" (UID: \"fde9c043-8307-4382-8ae1-6fd4114d938d\") " pod="openshift-must-gather-2bscz/crc-debug-ccpkj" Nov 24 07:52:07 crc kubenswrapper[4809]: I1124 07:52:07.328709 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fde9c043-8307-4382-8ae1-6fd4114d938d-host\") pod \"crc-debug-ccpkj\" (UID: \"fde9c043-8307-4382-8ae1-6fd4114d938d\") " pod="openshift-must-gather-2bscz/crc-debug-ccpkj" Nov 24 07:52:07 crc kubenswrapper[4809]: I1124 07:52:07.346309 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7tkm\" (UniqueName: \"kubernetes.io/projected/fde9c043-8307-4382-8ae1-6fd4114d938d-kube-api-access-f7tkm\") pod \"crc-debug-ccpkj\" (UID: \"fde9c043-8307-4382-8ae1-6fd4114d938d\") " pod="openshift-must-gather-2bscz/crc-debug-ccpkj" Nov 24 07:52:07 crc kubenswrapper[4809]: I1124 07:52:07.424805 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2bscz/crc-debug-ccpkj" Nov 24 07:52:07 crc kubenswrapper[4809]: W1124 07:52:07.450567 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfde9c043_8307_4382_8ae1_6fd4114d938d.slice/crio-8a76e87b351fe1736865d5e8c7db569f3e54a0b424e727dfa69ac1c3aafa00d7 WatchSource:0}: Error finding container 8a76e87b351fe1736865d5e8c7db569f3e54a0b424e727dfa69ac1c3aafa00d7: Status 404 returned error can't find the container with id 8a76e87b351fe1736865d5e8c7db569f3e54a0b424e727dfa69ac1c3aafa00d7 Nov 24 07:52:07 crc kubenswrapper[4809]: I1124 07:52:07.483204 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2bscz/crc-debug-6rtvd" Nov 24 07:52:07 crc kubenswrapper[4809]: I1124 07:52:07.483247 4809 scope.go:117] "RemoveContainer" containerID="42a03d852bb27ffacc27c7e4314c7008cf983f04875507399491c453d672a257" Nov 24 07:52:07 crc kubenswrapper[4809]: I1124 07:52:07.485851 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2bscz/crc-debug-ccpkj" event={"ID":"fde9c043-8307-4382-8ae1-6fd4114d938d","Type":"ContainerStarted","Data":"8a76e87b351fe1736865d5e8c7db569f3e54a0b424e727dfa69ac1c3aafa00d7"} Nov 24 07:52:08 crc kubenswrapper[4809]: I1124 07:52:08.496389 4809 generic.go:334] "Generic (PLEG): container finished" podID="fde9c043-8307-4382-8ae1-6fd4114d938d" containerID="75e69c0885839fa532b1a8a31c7bad228bc81b27ad918d32791f81034327464f" exitCode=0 Nov 24 07:52:08 crc kubenswrapper[4809]: I1124 07:52:08.496471 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2bscz/crc-debug-ccpkj" event={"ID":"fde9c043-8307-4382-8ae1-6fd4114d938d","Type":"ContainerDied","Data":"75e69c0885839fa532b1a8a31c7bad228bc81b27ad918d32791f81034327464f"} Nov 24 07:52:08 crc kubenswrapper[4809]: I1124 07:52:08.543570 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2bscz/crc-debug-ccpkj"] Nov 24 07:52:08 crc kubenswrapper[4809]: I1124 07:52:08.551315 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2bscz/crc-debug-ccpkj"] Nov 24 07:52:09 crc kubenswrapper[4809]: I1124 07:52:09.614486 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2bscz/crc-debug-ccpkj" Nov 24 07:52:09 crc kubenswrapper[4809]: I1124 07:52:09.680245 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fde9c043-8307-4382-8ae1-6fd4114d938d-host\") pod \"fde9c043-8307-4382-8ae1-6fd4114d938d\" (UID: \"fde9c043-8307-4382-8ae1-6fd4114d938d\") " Nov 24 07:52:09 crc kubenswrapper[4809]: I1124 07:52:09.680474 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fde9c043-8307-4382-8ae1-6fd4114d938d-host" (OuterVolumeSpecName: "host") pod "fde9c043-8307-4382-8ae1-6fd4114d938d" (UID: "fde9c043-8307-4382-8ae1-6fd4114d938d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:52:09 crc kubenswrapper[4809]: I1124 07:52:09.680504 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7tkm\" (UniqueName: \"kubernetes.io/projected/fde9c043-8307-4382-8ae1-6fd4114d938d-kube-api-access-f7tkm\") pod \"fde9c043-8307-4382-8ae1-6fd4114d938d\" (UID: \"fde9c043-8307-4382-8ae1-6fd4114d938d\") " Nov 24 07:52:09 crc kubenswrapper[4809]: I1124 07:52:09.681047 4809 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fde9c043-8307-4382-8ae1-6fd4114d938d-host\") on node \"crc\" DevicePath \"\"" Nov 24 07:52:09 crc kubenswrapper[4809]: I1124 07:52:09.686227 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fde9c043-8307-4382-8ae1-6fd4114d938d-kube-api-access-f7tkm" (OuterVolumeSpecName: "kube-api-access-f7tkm") pod "fde9c043-8307-4382-8ae1-6fd4114d938d" (UID: "fde9c043-8307-4382-8ae1-6fd4114d938d"). InnerVolumeSpecName "kube-api-access-f7tkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:52:09 crc kubenswrapper[4809]: I1124 07:52:09.782368 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7tkm\" (UniqueName: \"kubernetes.io/projected/fde9c043-8307-4382-8ae1-6fd4114d938d-kube-api-access-f7tkm\") on node \"crc\" DevicePath \"\"" Nov 24 07:52:10 crc kubenswrapper[4809]: I1124 07:52:10.525769 4809 scope.go:117] "RemoveContainer" containerID="75e69c0885839fa532b1a8a31c7bad228bc81b27ad918d32791f81034327464f" Nov 24 07:52:10 crc kubenswrapper[4809]: I1124 07:52:10.526391 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2bscz/crc-debug-ccpkj" Nov 24 07:52:10 crc kubenswrapper[4809]: I1124 07:52:10.900888 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fde9c043-8307-4382-8ae1-6fd4114d938d" path="/var/lib/kubelet/pods/fde9c043-8307-4382-8ae1-6fd4114d938d/volumes" Nov 24 07:52:22 crc kubenswrapper[4809]: I1124 07:52:22.963406 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7bb85696db-hj9dc_04780f37-736e-4f56-a887-e50c11e3b3ac/barbican-api/0.log" Nov 24 07:52:23 crc kubenswrapper[4809]: I1124 07:52:23.039510 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7bb85696db-hj9dc_04780f37-736e-4f56-a887-e50c11e3b3ac/barbican-api-log/0.log" Nov 24 07:52:23 crc kubenswrapper[4809]: I1124 07:52:23.283539 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7947c766c8-w6s92_ada052fc-743e-477f-b5be-848ca3536f09/barbican-keystone-listener/0.log" Nov 24 07:52:23 crc kubenswrapper[4809]: I1124 07:52:23.409450 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7947c766c8-w6s92_ada052fc-743e-477f-b5be-848ca3536f09/barbican-keystone-listener-log/0.log" Nov 24 07:52:23 crc kubenswrapper[4809]: I1124 07:52:23.457054 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-cc9bfd6b9-27jr8_5c05e3fc-9407-4bc4-86b4-409ebee7f3e7/barbican-worker/0.log" Nov 24 07:52:23 crc kubenswrapper[4809]: I1124 07:52:23.509513 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-cc9bfd6b9-27jr8_5c05e3fc-9407-4bc4-86b4-409ebee7f3e7/barbican-worker-log/0.log" Nov 24 07:52:23 crc kubenswrapper[4809]: I1124 07:52:23.641781 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28_bad47712-c44f-4799-bfe2-6c95ea3b513f/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:52:23 crc kubenswrapper[4809]: I1124 07:52:23.742281 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_96597dff-85f6-4758-a2e5-27c5b01bf34a/ceilometer-central-agent/0.log" Nov 24 07:52:23 crc kubenswrapper[4809]: I1124 07:52:23.808658 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_96597dff-85f6-4758-a2e5-27c5b01bf34a/ceilometer-notification-agent/0.log" Nov 24 07:52:23 crc kubenswrapper[4809]: I1124 07:52:23.839738 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_96597dff-85f6-4758-a2e5-27c5b01bf34a/proxy-httpd/0.log" Nov 24 07:52:23 crc kubenswrapper[4809]: I1124 07:52:23.936282 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_96597dff-85f6-4758-a2e5-27c5b01bf34a/sg-core/0.log" Nov 24 07:52:24 crc kubenswrapper[4809]: I1124 07:52:24.056456 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_097f9d2f-9e7a-42b8-a40e-da15a39b146c/cinder-api/0.log" Nov 24 07:52:24 crc kubenswrapper[4809]: I1124 07:52:24.079300 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_097f9d2f-9e7a-42b8-a40e-da15a39b146c/cinder-api-log/0.log" Nov 24 07:52:24 crc kubenswrapper[4809]: I1124 07:52:24.235449 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8ea4784f-e104-4ad0-873e-7d336cf4d98e/cinder-scheduler/0.log" Nov 24 07:52:24 crc kubenswrapper[4809]: I1124 07:52:24.279810 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8ea4784f-e104-4ad0-873e-7d336cf4d98e/probe/0.log" Nov 24 07:52:24 crc kubenswrapper[4809]: I1124 07:52:24.380157 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k_8a942ac4-7652-45aa-b238-3aff50848195/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:52:24 crc kubenswrapper[4809]: I1124 07:52:24.492344 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr_a7f9b89a-871e-4478-bb01-1120cd72a2b6/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:52:24 crc kubenswrapper[4809]: I1124 07:52:24.595127 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-f6c2x_09eb3227-9b7b-4d9a-9fa1-80e28696c61e/init/0.log" Nov 24 07:52:24 crc kubenswrapper[4809]: I1124 07:52:24.753237 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-f6c2x_09eb3227-9b7b-4d9a-9fa1-80e28696c61e/init/0.log" Nov 24 07:52:24 crc kubenswrapper[4809]: I1124 07:52:24.790627 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-f6c2x_09eb3227-9b7b-4d9a-9fa1-80e28696c61e/dnsmasq-dns/0.log" Nov 24 07:52:24 crc kubenswrapper[4809]: I1124 07:52:24.814775 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9_a022199a-9d09-421e-b9f3-67d00cab4f5f/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:52:24 crc kubenswrapper[4809]: I1124 07:52:24.961827 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_cfb0ea66-5650-4f85-a36a-06b60949a4bf/glance-httpd/0.log" Nov 24 07:52:25 crc kubenswrapper[4809]: I1124 07:52:25.015587 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_cfb0ea66-5650-4f85-a36a-06b60949a4bf/glance-log/0.log" Nov 24 07:52:25 crc kubenswrapper[4809]: I1124 07:52:25.153800 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_91d79979-625d-4ad3-bbb5-f3f35e7142a4/glance-log/0.log" Nov 24 07:52:25 crc kubenswrapper[4809]: I1124 07:52:25.210749 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_91d79979-625d-4ad3-bbb5-f3f35e7142a4/glance-httpd/0.log" Nov 24 07:52:25 crc kubenswrapper[4809]: I1124 07:52:25.407095 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-8585547888-qdlwv_b5487494-236a-4335-977f-11167eaa3cdc/horizon/0.log" Nov 24 07:52:25 crc kubenswrapper[4809]: I1124 07:52:25.528935 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9_d4e896d6-2d97-431e-bfaf-f48369c0af23/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:52:25 crc kubenswrapper[4809]: I1124 07:52:25.657512 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-8585547888-qdlwv_b5487494-236a-4335-977f-11167eaa3cdc/horizon-log/0.log" Nov 24 07:52:25 crc kubenswrapper[4809]: I1124 07:52:25.696301 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-2fc9j_241ce393-aede-4973-adcf-93efc73bcd32/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:52:25 crc kubenswrapper[4809]: I1124 07:52:25.874492 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_6f65b9e2-204d-455f-b8e1-43ccec98b7f6/kube-state-metrics/0.log" Nov 24 07:52:26 crc kubenswrapper[4809]: I1124 07:52:26.071380 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-fd8844f9c-6jj95_1a3f2bae-ab41-461b-87bc-455f850805e6/keystone-api/0.log" Nov 24 07:52:26 crc kubenswrapper[4809]: I1124 07:52:26.143520 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s_fd55a20b-bdd2-40bf-85d7-c0b485352b8f/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:52:26 crc kubenswrapper[4809]: I1124 07:52:26.514312 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6c5d67c78c-vvvv9_9c53f941-5f7d-489c-85eb-d07ff8d4ef6b/neutron-api/0.log" Nov 24 07:52:26 crc kubenswrapper[4809]: I1124 07:52:26.604549 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6c5d67c78c-vvvv9_9c53f941-5f7d-489c-85eb-d07ff8d4ef6b/neutron-httpd/0.log" Nov 24 07:52:26 crc kubenswrapper[4809]: I1124 07:52:26.722269 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x_aef096ab-d8a4-4b78-97be-17967897eb98/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:52:27 crc kubenswrapper[4809]: I1124 07:52:27.163767 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_663c08e1-fd97-42fd-b882-557ca9d71cdf/nova-cell0-conductor-conductor/0.log" Nov 24 07:52:27 crc kubenswrapper[4809]: I1124 07:52:27.173612 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_6e8e43ff-8f38-4981-a78c-41934ecc1210/nova-api-log/0.log" Nov 24 07:52:27 crc kubenswrapper[4809]: I1124 07:52:27.386708 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_6e8e43ff-8f38-4981-a78c-41934ecc1210/nova-api-api/0.log" Nov 24 07:52:27 crc kubenswrapper[4809]: I1124 07:52:27.473254 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_f4318b40-eb3b-4de1-8c8f-c9e4bf5996f8/nova-cell1-conductor-conductor/0.log" Nov 24 07:52:27 crc kubenswrapper[4809]: I1124 07:52:27.476300 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_c0f2a430-fd7f-419b-a07c-283144e6f0f2/nova-cell1-novncproxy-novncproxy/0.log" Nov 24 07:52:27 crc kubenswrapper[4809]: I1124 07:52:27.655128 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-wsk2q_4b9c2d10-ec36-41c7-985c-44ee750eddd0/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:52:27 crc kubenswrapper[4809]: I1124 07:52:27.762405 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_d9b4155b-3ac1-4c8d-b768-e28485799816/nova-metadata-log/0.log" Nov 24 07:52:28 crc kubenswrapper[4809]: I1124 07:52:28.138783 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_d4270af3-165e-49e5-b88e-1c07b1a3a950/nova-scheduler-scheduler/0.log" Nov 24 07:52:28 crc kubenswrapper[4809]: I1124 07:52:28.178828 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e5afba7d-dbf3-4aa6-863b-1d3a6d71af88/mysql-bootstrap/0.log" Nov 24 07:52:28 crc kubenswrapper[4809]: I1124 07:52:28.378932 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e5afba7d-dbf3-4aa6-863b-1d3a6d71af88/mysql-bootstrap/0.log" Nov 24 07:52:28 crc kubenswrapper[4809]: I1124 07:52:28.448756 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e5afba7d-dbf3-4aa6-863b-1d3a6d71af88/galera/0.log" Nov 24 07:52:28 crc kubenswrapper[4809]: I1124 07:52:28.652648 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_caea018d-04b5-4d20-8161-9a9a87f67583/mysql-bootstrap/0.log" Nov 24 07:52:28 crc kubenswrapper[4809]: I1124 07:52:28.769333 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_caea018d-04b5-4d20-8161-9a9a87f67583/mysql-bootstrap/0.log" Nov 24 07:52:28 crc kubenswrapper[4809]: I1124 07:52:28.779292 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_caea018d-04b5-4d20-8161-9a9a87f67583/galera/0.log" Nov 24 07:52:28 crc kubenswrapper[4809]: I1124 07:52:28.948566 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_45b7caaf-bebc-4bef-804b-a39246b436a0/openstackclient/0.log" Nov 24 07:52:28 crc kubenswrapper[4809]: I1124 07:52:28.950162 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_d9b4155b-3ac1-4c8d-b768-e28485799816/nova-metadata-metadata/0.log" Nov 24 07:52:29 crc kubenswrapper[4809]: I1124 07:52:29.098007 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-mp9kw_781916ac-bfd7-4ce4-85c0-1e938fae3103/openstack-network-exporter/0.log" Nov 24 07:52:29 crc kubenswrapper[4809]: I1124 07:52:29.188312 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-drz48_3c2d3d10-789d-4a2d-9926-cba915a127df/ovsdb-server-init/0.log" Nov 24 07:52:29 crc kubenswrapper[4809]: I1124 07:52:29.422999 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-drz48_3c2d3d10-789d-4a2d-9926-cba915a127df/ovs-vswitchd/0.log" Nov 24 07:52:29 crc kubenswrapper[4809]: I1124 07:52:29.423936 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-drz48_3c2d3d10-789d-4a2d-9926-cba915a127df/ovsdb-server/0.log" Nov 24 07:52:29 crc kubenswrapper[4809]: I1124 07:52:29.435669 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-drz48_3c2d3d10-789d-4a2d-9926-cba915a127df/ovsdb-server-init/0.log" Nov 24 07:52:29 crc kubenswrapper[4809]: I1124 07:52:29.698492 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-vnpwc_c14ae567-41ab-455a-8756-04daf7c81677/ovn-controller/0.log" Nov 24 07:52:29 crc kubenswrapper[4809]: I1124 07:52:29.705603 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-mfjkm_a25b9532-a2c1-4e47-b8c2-7c5cec0fa255/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:52:29 crc kubenswrapper[4809]: I1124 07:52:29.871121 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_acc130b5-9141-4d3a-aec1-7778963967b0/ovn-northd/0.log" Nov 24 07:52:29 crc kubenswrapper[4809]: I1124 07:52:29.878840 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_acc130b5-9141-4d3a-aec1-7778963967b0/openstack-network-exporter/0.log" Nov 24 07:52:29 crc kubenswrapper[4809]: I1124 07:52:29.970769 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4be21c61-834b-4cff-bae7-ec458991ac7c/openstack-network-exporter/0.log" Nov 24 07:52:30 crc kubenswrapper[4809]: I1124 07:52:30.132157 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe/openstack-network-exporter/0.log" Nov 24 07:52:30 crc kubenswrapper[4809]: I1124 07:52:30.132883 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4be21c61-834b-4cff-bae7-ec458991ac7c/ovsdbserver-nb/0.log" Nov 24 07:52:30 crc kubenswrapper[4809]: I1124 07:52:30.192871 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe/ovsdbserver-sb/0.log" Nov 24 07:52:30 crc kubenswrapper[4809]: I1124 07:52:30.370884 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-78fc658c7d-ntbgd_446a68c7-316d-43a1-af00-feba9b01e2dd/placement-api/0.log" Nov 24 07:52:30 crc kubenswrapper[4809]: I1124 07:52:30.395631 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-78fc658c7d-ntbgd_446a68c7-316d-43a1-af00-feba9b01e2dd/placement-log/0.log" Nov 24 07:52:30 crc kubenswrapper[4809]: I1124 07:52:30.520062 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3285fa33-509a-4959-8547-0332857328c9/setup-container/0.log" Nov 24 07:52:30 crc kubenswrapper[4809]: I1124 07:52:30.798838 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3285fa33-509a-4959-8547-0332857328c9/rabbitmq/0.log" Nov 24 07:52:30 crc kubenswrapper[4809]: I1124 07:52:30.824838 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3285fa33-509a-4959-8547-0332857328c9/setup-container/0.log" Nov 24 07:52:30 crc kubenswrapper[4809]: I1124 07:52:30.873301 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_4a1ec853-3d6a-403a-b2fe-82dc89ac5779/setup-container/0.log" Nov 24 07:52:31 crc kubenswrapper[4809]: I1124 07:52:31.092583 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2_a46047d8-9af2-4172-8e55-26dd29b587aa/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:52:31 crc kubenswrapper[4809]: I1124 07:52:31.147297 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_4a1ec853-3d6a-403a-b2fe-82dc89ac5779/rabbitmq/0.log" Nov 24 07:52:31 crc kubenswrapper[4809]: I1124 07:52:31.149073 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_4a1ec853-3d6a-403a-b2fe-82dc89ac5779/setup-container/0.log" Nov 24 07:52:31 crc kubenswrapper[4809]: I1124 07:52:31.337318 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-wmc74_7883d395-4314-4e1c-9a89-89b949d1b7a6/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:52:31 crc kubenswrapper[4809]: I1124 07:52:31.392105 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9_9ed9735b-66b5-44b3-9399-af938bfcd802/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:52:31 crc kubenswrapper[4809]: I1124 07:52:31.595147 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-2fc45_2a469d7b-c3d8-46cf-b6cf-7fe13619f922/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:52:31 crc kubenswrapper[4809]: I1124 07:52:31.664626 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-ckss4_801a6c69-8f03-4f88-ba22-cd68875e8ae2/ssh-known-hosts-edpm-deployment/0.log" Nov 24 07:52:31 crc kubenswrapper[4809]: I1124 07:52:31.878159 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6d74d65d8c-rgmvx_2c1fe2a7-0b4f-46ee-8368-624499f8c095/proxy-server/0.log" Nov 24 07:52:31 crc kubenswrapper[4809]: I1124 07:52:31.935993 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6d74d65d8c-rgmvx_2c1fe2a7-0b4f-46ee-8368-624499f8c095/proxy-httpd/0.log" Nov 24 07:52:31 crc kubenswrapper[4809]: I1124 07:52:31.984034 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-8n9sj_9ad3cf13-9872-497f-b8ee-180ec8d7bbdd/swift-ring-rebalance/0.log" Nov 24 07:52:32 crc kubenswrapper[4809]: I1124 07:52:32.152638 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/account-auditor/0.log" Nov 24 07:52:32 crc kubenswrapper[4809]: I1124 07:52:32.167903 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/account-reaper/0.log" Nov 24 07:52:32 crc kubenswrapper[4809]: I1124 07:52:32.275170 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/account-replicator/0.log" Nov 24 07:52:32 crc kubenswrapper[4809]: I1124 07:52:32.315006 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/account-server/0.log" Nov 24 07:52:32 crc kubenswrapper[4809]: I1124 07:52:32.382942 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/container-auditor/0.log" Nov 24 07:52:32 crc kubenswrapper[4809]: I1124 07:52:32.418117 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/container-replicator/0.log" Nov 24 07:52:32 crc kubenswrapper[4809]: I1124 07:52:32.487692 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/container-server/0.log" Nov 24 07:52:32 crc kubenswrapper[4809]: I1124 07:52:32.562664 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/container-updater/0.log" Nov 24 07:52:32 crc kubenswrapper[4809]: I1124 07:52:32.580947 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/object-auditor/0.log" Nov 24 07:52:32 crc kubenswrapper[4809]: I1124 07:52:32.643123 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/object-expirer/0.log" Nov 24 07:52:32 crc kubenswrapper[4809]: I1124 07:52:32.756107 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/object-replicator/0.log" Nov 24 07:52:32 crc kubenswrapper[4809]: I1124 07:52:32.776657 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/object-server/0.log" Nov 24 07:52:32 crc kubenswrapper[4809]: I1124 07:52:32.826576 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/object-updater/0.log" Nov 24 07:52:32 crc kubenswrapper[4809]: I1124 07:52:32.879228 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/rsync/0.log" Nov 24 07:52:32 crc kubenswrapper[4809]: I1124 07:52:32.971059 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/swift-recon-cron/0.log" Nov 24 07:52:33 crc kubenswrapper[4809]: I1124 07:52:33.132135 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd_797b01f0-7757-495a-ad23-3973353832c1/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:52:33 crc kubenswrapper[4809]: I1124 07:52:33.174994 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_4f90573b-06dc-47d5-ab8b-5dab5503e96a/tempest-tests-tempest-tests-runner/0.log" Nov 24 07:52:33 crc kubenswrapper[4809]: I1124 07:52:33.283496 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_19e09f12-9159-4150-ae5b-6d718e6f9957/test-operator-logs-container/0.log" Nov 24 07:52:33 crc kubenswrapper[4809]: I1124 07:52:33.439713 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f_2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:52:41 crc kubenswrapper[4809]: I1124 07:52:41.988875 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_ced578a4-f9b1-4d1f-978f-beda4b33bc59/memcached/0.log" Nov 24 07:52:55 crc kubenswrapper[4809]: I1124 07:52:55.026158 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b_ed47ae85-6bd2-4e58-acee-30c9022b9170/util/0.log" Nov 24 07:52:55 crc kubenswrapper[4809]: I1124 07:52:55.159933 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b_ed47ae85-6bd2-4e58-acee-30c9022b9170/pull/0.log" Nov 24 07:52:55 crc kubenswrapper[4809]: I1124 07:52:55.181913 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b_ed47ae85-6bd2-4e58-acee-30c9022b9170/util/0.log" Nov 24 07:52:55 crc kubenswrapper[4809]: I1124 07:52:55.199577 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b_ed47ae85-6bd2-4e58-acee-30c9022b9170/pull/0.log" Nov 24 07:52:55 crc kubenswrapper[4809]: I1124 07:52:55.388562 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b_ed47ae85-6bd2-4e58-acee-30c9022b9170/extract/0.log" Nov 24 07:52:55 crc kubenswrapper[4809]: I1124 07:52:55.420870 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b_ed47ae85-6bd2-4e58-acee-30c9022b9170/pull/0.log" Nov 24 07:52:55 crc kubenswrapper[4809]: I1124 07:52:55.454351 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b_ed47ae85-6bd2-4e58-acee-30c9022b9170/util/0.log" Nov 24 07:52:55 crc kubenswrapper[4809]: I1124 07:52:55.551850 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-6k9xr_bfcf902b-83f0-48f9-ad6b-eb7963f43af4/kube-rbac-proxy/0.log" Nov 24 07:52:55 crc kubenswrapper[4809]: I1124 07:52:55.633850 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-49dpw_e525562e-b6b2-415f-b6e7-1c0bfde8a56f/kube-rbac-proxy/0.log" Nov 24 07:52:55 crc kubenswrapper[4809]: I1124 07:52:55.648204 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-6k9xr_bfcf902b-83f0-48f9-ad6b-eb7963f43af4/manager/0.log" Nov 24 07:52:55 crc kubenswrapper[4809]: I1124 07:52:55.790527 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-49dpw_e525562e-b6b2-415f-b6e7-1c0bfde8a56f/manager/0.log" Nov 24 07:52:55 crc kubenswrapper[4809]: I1124 07:52:55.843566 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-plhlb_200f6896-3100-46ea-94f4-69637762a259/kube-rbac-proxy/0.log" Nov 24 07:52:55 crc kubenswrapper[4809]: I1124 07:52:55.868994 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-plhlb_200f6896-3100-46ea-94f4-69637762a259/manager/0.log" Nov 24 07:52:55 crc kubenswrapper[4809]: I1124 07:52:55.993581 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-8nzmk_0a1424d6-9ecf-4a06-b174-15d13f61b916/kube-rbac-proxy/0.log" Nov 24 07:52:56 crc kubenswrapper[4809]: I1124 07:52:56.109880 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-8nzmk_0a1424d6-9ecf-4a06-b174-15d13f61b916/manager/0.log" Nov 24 07:52:56 crc kubenswrapper[4809]: I1124 07:52:56.182040 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-4ddch_d78cc4fe-bab9-4521-a1d2-ed6efd341789/kube-rbac-proxy/0.log" Nov 24 07:52:56 crc kubenswrapper[4809]: I1124 07:52:56.197266 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-4ddch_d78cc4fe-bab9-4521-a1d2-ed6efd341789/manager/0.log" Nov 24 07:52:56 crc kubenswrapper[4809]: I1124 07:52:56.282827 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-xf4fh_4d4aa35c-faa7-4b93-aecd-ee4645f11916/kube-rbac-proxy/0.log" Nov 24 07:52:56 crc kubenswrapper[4809]: I1124 07:52:56.376474 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-xf4fh_4d4aa35c-faa7-4b93-aecd-ee4645f11916/manager/0.log" Nov 24 07:52:56 crc kubenswrapper[4809]: I1124 07:52:56.447375 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-4vrfb_5b90cc28-7c58-4826-b849-3456bb654f0d/kube-rbac-proxy/0.log" Nov 24 07:52:56 crc kubenswrapper[4809]: I1124 07:52:56.622518 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-jfjqt_5f4f9ac0-6714-4f4a-a7fb-8d65b19c919c/kube-rbac-proxy/0.log" Nov 24 07:52:56 crc kubenswrapper[4809]: I1124 07:52:56.651780 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-jfjqt_5f4f9ac0-6714-4f4a-a7fb-8d65b19c919c/manager/0.log" Nov 24 07:52:56 crc kubenswrapper[4809]: I1124 07:52:56.661267 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-4vrfb_5b90cc28-7c58-4826-b849-3456bb654f0d/manager/0.log" Nov 24 07:52:56 crc kubenswrapper[4809]: I1124 07:52:56.814024 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-wh5fq_b0a19ac9-d266-4416-8590-3c8de4dadc7f/kube-rbac-proxy/0.log" Nov 24 07:52:56 crc kubenswrapper[4809]: I1124 07:52:56.877697 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-wh5fq_b0a19ac9-d266-4416-8590-3c8de4dadc7f/manager/0.log" Nov 24 07:52:56 crc kubenswrapper[4809]: I1124 07:52:56.963293 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-bz8c8_a2b969b3-07c7-4bf3-ab18-5317d8f0a372/kube-rbac-proxy/0.log" Nov 24 07:52:57 crc kubenswrapper[4809]: I1124 07:52:57.011385 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-bz8c8_a2b969b3-07c7-4bf3-ab18-5317d8f0a372/manager/0.log" Nov 24 07:52:57 crc kubenswrapper[4809]: I1124 07:52:57.055506 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-rp7r4_2f364be5-58b1-4d57-97f5-65a13d63e34e/kube-rbac-proxy/0.log" Nov 24 07:52:57 crc kubenswrapper[4809]: I1124 07:52:57.151654 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-rp7r4_2f364be5-58b1-4d57-97f5-65a13d63e34e/manager/0.log" Nov 24 07:52:57 crc kubenswrapper[4809]: I1124 07:52:57.232241 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-rgkd9_e94d989e-7aee-43d9-87c0-712355ca193f/kube-rbac-proxy/0.log" Nov 24 07:52:57 crc kubenswrapper[4809]: I1124 07:52:57.312038 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-rgkd9_e94d989e-7aee-43d9-87c0-712355ca193f/manager/0.log" Nov 24 07:52:57 crc kubenswrapper[4809]: I1124 07:52:57.381269 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-tk2hk_fd9ddd03-bc95-43e1-93ef-97ad5ca589b3/kube-rbac-proxy/0.log" Nov 24 07:52:57 crc kubenswrapper[4809]: I1124 07:52:57.531837 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-tk2hk_fd9ddd03-bc95-43e1-93ef-97ad5ca589b3/manager/0.log" Nov 24 07:52:57 crc kubenswrapper[4809]: I1124 07:52:57.563421 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-84nbm_bfe0f53e-685b-4cef-b763-90399c5fdaac/kube-rbac-proxy/0.log" Nov 24 07:52:57 crc kubenswrapper[4809]: I1124 07:52:57.600855 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-84nbm_bfe0f53e-685b-4cef-b763-90399c5fdaac/manager/0.log" Nov 24 07:52:57 crc kubenswrapper[4809]: I1124 07:52:57.747859 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-7vg75_2e599303-b44b-4ce5-9ee9-1e1c13ca856f/kube-rbac-proxy/0.log" Nov 24 07:52:57 crc kubenswrapper[4809]: I1124 07:52:57.748138 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-7vg75_2e599303-b44b-4ce5-9ee9-1e1c13ca856f/manager/0.log" Nov 24 07:52:57 crc kubenswrapper[4809]: I1124 07:52:57.976269 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8fcb84d5c-xqd67_5b38e6fb-1aef-4ecd-80ca-4665a9d3430a/kube-rbac-proxy/0.log" Nov 24 07:52:58 crc kubenswrapper[4809]: I1124 07:52:58.059949 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6c94b45f4d-pdjb4_a9cc582d-d6ce-4bb6-b698-9733ce8be925/kube-rbac-proxy/0.log" Nov 24 07:52:58 crc kubenswrapper[4809]: I1124 07:52:58.290361 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-mhj4l_bb31150b-b8eb-4340-ba08-384de196289a/registry-server/0.log" Nov 24 07:52:58 crc kubenswrapper[4809]: I1124 07:52:58.436646 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6c94b45f4d-pdjb4_a9cc582d-d6ce-4bb6-b698-9733ce8be925/operator/0.log" Nov 24 07:52:58 crc kubenswrapper[4809]: I1124 07:52:58.524781 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-59brn_9eeb9814-10f8-478d-888f-5cf058c99c4b/kube-rbac-proxy/0.log" Nov 24 07:52:58 crc kubenswrapper[4809]: I1124 07:52:58.767176 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-rtxsf_483b5843-7751-465a-9412-dbb15ef5145f/kube-rbac-proxy/0.log" Nov 24 07:52:58 crc kubenswrapper[4809]: I1124 07:52:58.782328 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-59brn_9eeb9814-10f8-478d-888f-5cf058c99c4b/manager/0.log" Nov 24 07:52:59 crc kubenswrapper[4809]: I1124 07:52:59.031888 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-px9vl_02a1aa7c-c3f7-4dcb-89ad-8d34f46ff4a1/operator/0.log" Nov 24 07:52:59 crc kubenswrapper[4809]: I1124 07:52:59.073308 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-rtxsf_483b5843-7751-465a-9412-dbb15ef5145f/manager/0.log" Nov 24 07:52:59 crc kubenswrapper[4809]: I1124 07:52:59.176801 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8fcb84d5c-xqd67_5b38e6fb-1aef-4ecd-80ca-4665a9d3430a/manager/0.log" Nov 24 07:52:59 crc kubenswrapper[4809]: I1124 07:52:59.313120 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-mfgld_6269a5dc-fd2a-4fdd-9b18-4e505242efca/kube-rbac-proxy/0.log" Nov 24 07:52:59 crc kubenswrapper[4809]: I1124 07:52:59.318616 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-mfgld_6269a5dc-fd2a-4fdd-9b18-4e505242efca/manager/0.log" Nov 24 07:52:59 crc kubenswrapper[4809]: I1124 07:52:59.335111 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6d4bf84b58-4xjsj_39fc255b-7ac9-40b4-98e0-74d1f76c2696/kube-rbac-proxy/0.log" Nov 24 07:52:59 crc kubenswrapper[4809]: I1124 07:52:59.470171 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6d4bf84b58-4xjsj_39fc255b-7ac9-40b4-98e0-74d1f76c2696/manager/0.log" Nov 24 07:52:59 crc kubenswrapper[4809]: I1124 07:52:59.530437 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-wbzcn_cb240075-d846-4751-b026-6fa452a565a5/manager/0.log" Nov 24 07:52:59 crc kubenswrapper[4809]: I1124 07:52:59.561353 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-wbzcn_cb240075-d846-4751-b026-6fa452a565a5/kube-rbac-proxy/0.log" Nov 24 07:52:59 crc kubenswrapper[4809]: I1124 07:52:59.717218 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-twjz4_df289b0d-b465-49d2-b1ee-4f0faf87b079/kube-rbac-proxy/0.log" Nov 24 07:52:59 crc kubenswrapper[4809]: I1124 07:52:59.738517 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-twjz4_df289b0d-b465-49d2-b1ee-4f0faf87b079/manager/0.log" Nov 24 07:53:16 crc kubenswrapper[4809]: I1124 07:53:16.429755 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-xfdvj_3e14d56e-43b3-40ec-809f-371e8ab2f628/control-plane-machine-set-operator/0.log" Nov 24 07:53:16 crc kubenswrapper[4809]: I1124 07:53:16.576225 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-56j57_eca6947b-dd5e-449d-a8c5-05277c556bd1/kube-rbac-proxy/0.log" Nov 24 07:53:16 crc kubenswrapper[4809]: I1124 07:53:16.655077 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-56j57_eca6947b-dd5e-449d-a8c5-05277c556bd1/machine-api-operator/0.log" Nov 24 07:53:28 crc kubenswrapper[4809]: I1124 07:53:28.476763 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-hm272_1b39cb48-19bf-4ff6-a9e7-a2270b28ab87/cert-manager-controller/0.log" Nov 24 07:53:28 crc kubenswrapper[4809]: I1124 07:53:28.647995 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-xhsz2_9de96055-2c3b-4012-a260-7b55cbb0df1b/cert-manager-cainjector/0.log" Nov 24 07:53:28 crc kubenswrapper[4809]: I1124 07:53:28.771642 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-wslw7_4e03214c-a15c-4a2f-a802-ce80e9fde1c8/cert-manager-webhook/0.log" Nov 24 07:53:36 crc kubenswrapper[4809]: I1124 07:53:36.707732 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bmn4k"] Nov 24 07:53:36 crc kubenswrapper[4809]: E1124 07:53:36.711132 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fde9c043-8307-4382-8ae1-6fd4114d938d" containerName="container-00" Nov 24 07:53:36 crc kubenswrapper[4809]: I1124 07:53:36.711250 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fde9c043-8307-4382-8ae1-6fd4114d938d" containerName="container-00" Nov 24 07:53:36 crc kubenswrapper[4809]: I1124 07:53:36.711697 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fde9c043-8307-4382-8ae1-6fd4114d938d" containerName="container-00" Nov 24 07:53:36 crc kubenswrapper[4809]: I1124 07:53:36.722792 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bmn4k" Nov 24 07:53:36 crc kubenswrapper[4809]: I1124 07:53:36.724808 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bmn4k"] Nov 24 07:53:36 crc kubenswrapper[4809]: I1124 07:53:36.770309 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17f65a89-36a1-48aa-b66e-c45b15bfbaf0-catalog-content\") pod \"redhat-marketplace-bmn4k\" (UID: \"17f65a89-36a1-48aa-b66e-c45b15bfbaf0\") " pod="openshift-marketplace/redhat-marketplace-bmn4k" Nov 24 07:53:36 crc kubenswrapper[4809]: I1124 07:53:36.770385 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vt5mv\" (UniqueName: \"kubernetes.io/projected/17f65a89-36a1-48aa-b66e-c45b15bfbaf0-kube-api-access-vt5mv\") pod \"redhat-marketplace-bmn4k\" (UID: \"17f65a89-36a1-48aa-b66e-c45b15bfbaf0\") " pod="openshift-marketplace/redhat-marketplace-bmn4k" Nov 24 07:53:36 crc kubenswrapper[4809]: I1124 07:53:36.770565 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17f65a89-36a1-48aa-b66e-c45b15bfbaf0-utilities\") pod \"redhat-marketplace-bmn4k\" (UID: \"17f65a89-36a1-48aa-b66e-c45b15bfbaf0\") " pod="openshift-marketplace/redhat-marketplace-bmn4k" Nov 24 07:53:36 crc kubenswrapper[4809]: I1124 07:53:36.873524 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17f65a89-36a1-48aa-b66e-c45b15bfbaf0-utilities\") pod \"redhat-marketplace-bmn4k\" (UID: \"17f65a89-36a1-48aa-b66e-c45b15bfbaf0\") " pod="openshift-marketplace/redhat-marketplace-bmn4k" Nov 24 07:53:36 crc kubenswrapper[4809]: I1124 07:53:36.873708 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17f65a89-36a1-48aa-b66e-c45b15bfbaf0-catalog-content\") pod \"redhat-marketplace-bmn4k\" (UID: \"17f65a89-36a1-48aa-b66e-c45b15bfbaf0\") " pod="openshift-marketplace/redhat-marketplace-bmn4k" Nov 24 07:53:36 crc kubenswrapper[4809]: I1124 07:53:36.873740 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vt5mv\" (UniqueName: \"kubernetes.io/projected/17f65a89-36a1-48aa-b66e-c45b15bfbaf0-kube-api-access-vt5mv\") pod \"redhat-marketplace-bmn4k\" (UID: \"17f65a89-36a1-48aa-b66e-c45b15bfbaf0\") " pod="openshift-marketplace/redhat-marketplace-bmn4k" Nov 24 07:53:36 crc kubenswrapper[4809]: I1124 07:53:36.874308 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17f65a89-36a1-48aa-b66e-c45b15bfbaf0-utilities\") pod \"redhat-marketplace-bmn4k\" (UID: \"17f65a89-36a1-48aa-b66e-c45b15bfbaf0\") " pod="openshift-marketplace/redhat-marketplace-bmn4k" Nov 24 07:53:36 crc kubenswrapper[4809]: I1124 07:53:36.874428 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17f65a89-36a1-48aa-b66e-c45b15bfbaf0-catalog-content\") pod \"redhat-marketplace-bmn4k\" (UID: \"17f65a89-36a1-48aa-b66e-c45b15bfbaf0\") " pod="openshift-marketplace/redhat-marketplace-bmn4k" Nov 24 07:53:36 crc kubenswrapper[4809]: I1124 07:53:36.905050 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vt5mv\" (UniqueName: \"kubernetes.io/projected/17f65a89-36a1-48aa-b66e-c45b15bfbaf0-kube-api-access-vt5mv\") pod \"redhat-marketplace-bmn4k\" (UID: \"17f65a89-36a1-48aa-b66e-c45b15bfbaf0\") " pod="openshift-marketplace/redhat-marketplace-bmn4k" Nov 24 07:53:37 crc kubenswrapper[4809]: I1124 07:53:37.057570 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bmn4k" Nov 24 07:53:37 crc kubenswrapper[4809]: I1124 07:53:37.601954 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bmn4k"] Nov 24 07:53:38 crc kubenswrapper[4809]: I1124 07:53:38.374500 4809 generic.go:334] "Generic (PLEG): container finished" podID="17f65a89-36a1-48aa-b66e-c45b15bfbaf0" containerID="7de3fcf2ff39acf4f061e814a3f48e4b2d911a557b5ff7cd817a8a2e0b1d80d7" exitCode=0 Nov 24 07:53:38 crc kubenswrapper[4809]: I1124 07:53:38.374630 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bmn4k" event={"ID":"17f65a89-36a1-48aa-b66e-c45b15bfbaf0","Type":"ContainerDied","Data":"7de3fcf2ff39acf4f061e814a3f48e4b2d911a557b5ff7cd817a8a2e0b1d80d7"} Nov 24 07:53:38 crc kubenswrapper[4809]: I1124 07:53:38.375392 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bmn4k" event={"ID":"17f65a89-36a1-48aa-b66e-c45b15bfbaf0","Type":"ContainerStarted","Data":"213e9b0b805c24f2e449ac9d33681656d12cbd66cdcf8e6d61d1c48a819f470c"} Nov 24 07:53:39 crc kubenswrapper[4809]: I1124 07:53:39.304667 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zxwwc"] Nov 24 07:53:39 crc kubenswrapper[4809]: I1124 07:53:39.314384 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zxwwc" Nov 24 07:53:39 crc kubenswrapper[4809]: I1124 07:53:39.325100 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zxwwc"] Nov 24 07:53:39 crc kubenswrapper[4809]: I1124 07:53:39.449545 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrl4j\" (UniqueName: \"kubernetes.io/projected/7dd5163d-c45a-4fb9-a369-aef70a9739e9-kube-api-access-lrl4j\") pod \"community-operators-zxwwc\" (UID: \"7dd5163d-c45a-4fb9-a369-aef70a9739e9\") " pod="openshift-marketplace/community-operators-zxwwc" Nov 24 07:53:39 crc kubenswrapper[4809]: I1124 07:53:39.449655 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dd5163d-c45a-4fb9-a369-aef70a9739e9-utilities\") pod \"community-operators-zxwwc\" (UID: \"7dd5163d-c45a-4fb9-a369-aef70a9739e9\") " pod="openshift-marketplace/community-operators-zxwwc" Nov 24 07:53:39 crc kubenswrapper[4809]: I1124 07:53:39.450268 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dd5163d-c45a-4fb9-a369-aef70a9739e9-catalog-content\") pod \"community-operators-zxwwc\" (UID: \"7dd5163d-c45a-4fb9-a369-aef70a9739e9\") " pod="openshift-marketplace/community-operators-zxwwc" Nov 24 07:53:39 crc kubenswrapper[4809]: I1124 07:53:39.554002 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrl4j\" (UniqueName: \"kubernetes.io/projected/7dd5163d-c45a-4fb9-a369-aef70a9739e9-kube-api-access-lrl4j\") pod \"community-operators-zxwwc\" (UID: \"7dd5163d-c45a-4fb9-a369-aef70a9739e9\") " pod="openshift-marketplace/community-operators-zxwwc" Nov 24 07:53:39 crc kubenswrapper[4809]: I1124 07:53:39.554124 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dd5163d-c45a-4fb9-a369-aef70a9739e9-utilities\") pod \"community-operators-zxwwc\" (UID: \"7dd5163d-c45a-4fb9-a369-aef70a9739e9\") " pod="openshift-marketplace/community-operators-zxwwc" Nov 24 07:53:39 crc kubenswrapper[4809]: I1124 07:53:39.554251 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dd5163d-c45a-4fb9-a369-aef70a9739e9-catalog-content\") pod \"community-operators-zxwwc\" (UID: \"7dd5163d-c45a-4fb9-a369-aef70a9739e9\") " pod="openshift-marketplace/community-operators-zxwwc" Nov 24 07:53:39 crc kubenswrapper[4809]: I1124 07:53:39.554891 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dd5163d-c45a-4fb9-a369-aef70a9739e9-utilities\") pod \"community-operators-zxwwc\" (UID: \"7dd5163d-c45a-4fb9-a369-aef70a9739e9\") " pod="openshift-marketplace/community-operators-zxwwc" Nov 24 07:53:39 crc kubenswrapper[4809]: I1124 07:53:39.554913 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dd5163d-c45a-4fb9-a369-aef70a9739e9-catalog-content\") pod \"community-operators-zxwwc\" (UID: \"7dd5163d-c45a-4fb9-a369-aef70a9739e9\") " pod="openshift-marketplace/community-operators-zxwwc" Nov 24 07:53:39 crc kubenswrapper[4809]: I1124 07:53:39.584508 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrl4j\" (UniqueName: \"kubernetes.io/projected/7dd5163d-c45a-4fb9-a369-aef70a9739e9-kube-api-access-lrl4j\") pod \"community-operators-zxwwc\" (UID: \"7dd5163d-c45a-4fb9-a369-aef70a9739e9\") " pod="openshift-marketplace/community-operators-zxwwc" Nov 24 07:53:39 crc kubenswrapper[4809]: I1124 07:53:39.646476 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zxwwc" Nov 24 07:53:40 crc kubenswrapper[4809]: I1124 07:53:40.273133 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zxwwc"] Nov 24 07:53:40 crc kubenswrapper[4809]: I1124 07:53:40.406666 4809 generic.go:334] "Generic (PLEG): container finished" podID="17f65a89-36a1-48aa-b66e-c45b15bfbaf0" containerID="a6fcff21dec9f2ea855b7120977cbc1854a4880b1f3e8cb79fab8ce40ddf4ac8" exitCode=0 Nov 24 07:53:40 crc kubenswrapper[4809]: I1124 07:53:40.406800 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bmn4k" event={"ID":"17f65a89-36a1-48aa-b66e-c45b15bfbaf0","Type":"ContainerDied","Data":"a6fcff21dec9f2ea855b7120977cbc1854a4880b1f3e8cb79fab8ce40ddf4ac8"} Nov 24 07:53:40 crc kubenswrapper[4809]: I1124 07:53:40.408304 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxwwc" event={"ID":"7dd5163d-c45a-4fb9-a369-aef70a9739e9","Type":"ContainerStarted","Data":"1c84d19ddcd6e41c0ce4dc4d629b634739a08ff4b03e1fce6ab001d60d65d1a8"} Nov 24 07:53:41 crc kubenswrapper[4809]: I1124 07:53:41.419366 4809 generic.go:334] "Generic (PLEG): container finished" podID="7dd5163d-c45a-4fb9-a369-aef70a9739e9" containerID="7862ee532a8042ac7e5cd0e2f0bcfab5881ff10710a32b373ece5a546f3778b6" exitCode=0 Nov 24 07:53:41 crc kubenswrapper[4809]: I1124 07:53:41.419503 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxwwc" event={"ID":"7dd5163d-c45a-4fb9-a369-aef70a9739e9","Type":"ContainerDied","Data":"7862ee532a8042ac7e5cd0e2f0bcfab5881ff10710a32b373ece5a546f3778b6"} Nov 24 07:53:41 crc kubenswrapper[4809]: I1124 07:53:41.422178 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bmn4k" event={"ID":"17f65a89-36a1-48aa-b66e-c45b15bfbaf0","Type":"ContainerStarted","Data":"950dc8d1ea7e82bd660d42d4d4baee8493b79dccfa2400a3babadee502e0e420"} Nov 24 07:53:41 crc kubenswrapper[4809]: I1124 07:53:41.471894 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bmn4k" podStartSLOduration=3.052900469 podStartE2EDuration="5.471865883s" podCreationTimestamp="2025-11-24 07:53:36 +0000 UTC" firstStartedPulling="2025-11-24 07:53:38.378734241 +0000 UTC m=+3518.279325956" lastFinishedPulling="2025-11-24 07:53:40.797699675 +0000 UTC m=+3520.698291370" observedRunningTime="2025-11-24 07:53:41.461426222 +0000 UTC m=+3521.362017927" watchObservedRunningTime="2025-11-24 07:53:41.471865883 +0000 UTC m=+3521.372457588" Nov 24 07:53:42 crc kubenswrapper[4809]: I1124 07:53:42.324284 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-8fgt8_7265684f-a24a-40d3-8eb5-eb2da39cb870/nmstate-console-plugin/0.log" Nov 24 07:53:42 crc kubenswrapper[4809]: I1124 07:53:42.435462 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxwwc" event={"ID":"7dd5163d-c45a-4fb9-a369-aef70a9739e9","Type":"ContainerStarted","Data":"a8bcdf784693d98e9d26622ae75814814ded29c657a773b229a155f52ca7caea"} Nov 24 07:53:42 crc kubenswrapper[4809]: I1124 07:53:42.481808 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-5dl97_ef4d3282-be25-4395-899d-92b0fc8a0a40/nmstate-handler/0.log" Nov 24 07:53:42 crc kubenswrapper[4809]: I1124 07:53:42.541510 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-782jg_c00579c8-50fc-4b4e-8ddf-79ccdc0647ec/kube-rbac-proxy/0.log" Nov 24 07:53:42 crc kubenswrapper[4809]: I1124 07:53:42.596253 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-782jg_c00579c8-50fc-4b4e-8ddf-79ccdc0647ec/nmstate-metrics/0.log" Nov 24 07:53:42 crc kubenswrapper[4809]: I1124 07:53:42.766915 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-hbbwq_dec4efb0-1f05-416f-8e23-773aaab020df/nmstate-operator/0.log" Nov 24 07:53:42 crc kubenswrapper[4809]: I1124 07:53:42.866658 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-7fhk7_5c12932d-5d71-4a4f-9dc5-bbf20584f029/nmstate-webhook/0.log" Nov 24 07:53:44 crc kubenswrapper[4809]: I1124 07:53:44.457333 4809 generic.go:334] "Generic (PLEG): container finished" podID="7dd5163d-c45a-4fb9-a369-aef70a9739e9" containerID="a8bcdf784693d98e9d26622ae75814814ded29c657a773b229a155f52ca7caea" exitCode=0 Nov 24 07:53:44 crc kubenswrapper[4809]: I1124 07:53:44.457480 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxwwc" event={"ID":"7dd5163d-c45a-4fb9-a369-aef70a9739e9","Type":"ContainerDied","Data":"a8bcdf784693d98e9d26622ae75814814ded29c657a773b229a155f52ca7caea"} Nov 24 07:53:45 crc kubenswrapper[4809]: I1124 07:53:45.469056 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxwwc" event={"ID":"7dd5163d-c45a-4fb9-a369-aef70a9739e9","Type":"ContainerStarted","Data":"a7e7552aafe46e8ab3d4c5f8f1b25e1cbf7b2104ecd7f92c3af014709e60d7e9"} Nov 24 07:53:45 crc kubenswrapper[4809]: I1124 07:53:45.498287 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zxwwc" podStartSLOduration=3.038821624 podStartE2EDuration="6.498251027s" podCreationTimestamp="2025-11-24 07:53:39 +0000 UTC" firstStartedPulling="2025-11-24 07:53:41.421912488 +0000 UTC m=+3521.322504193" lastFinishedPulling="2025-11-24 07:53:44.881341891 +0000 UTC m=+3524.781933596" observedRunningTime="2025-11-24 07:53:45.490724415 +0000 UTC m=+3525.391316120" watchObservedRunningTime="2025-11-24 07:53:45.498251027 +0000 UTC m=+3525.398842732" Nov 24 07:53:47 crc kubenswrapper[4809]: I1124 07:53:47.058100 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bmn4k" Nov 24 07:53:47 crc kubenswrapper[4809]: I1124 07:53:47.058680 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bmn4k" Nov 24 07:53:47 crc kubenswrapper[4809]: I1124 07:53:47.116585 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bmn4k" Nov 24 07:53:47 crc kubenswrapper[4809]: I1124 07:53:47.547582 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bmn4k" Nov 24 07:53:48 crc kubenswrapper[4809]: I1124 07:53:48.043448 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:53:48 crc kubenswrapper[4809]: I1124 07:53:48.043531 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:53:48 crc kubenswrapper[4809]: I1124 07:53:48.285014 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bmn4k"] Nov 24 07:53:49 crc kubenswrapper[4809]: I1124 07:53:49.505811 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bmn4k" podUID="17f65a89-36a1-48aa-b66e-c45b15bfbaf0" containerName="registry-server" containerID="cri-o://950dc8d1ea7e82bd660d42d4d4baee8493b79dccfa2400a3babadee502e0e420" gracePeriod=2 Nov 24 07:53:49 crc kubenswrapper[4809]: I1124 07:53:49.647865 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zxwwc" Nov 24 07:53:49 crc kubenswrapper[4809]: I1124 07:53:49.649068 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zxwwc" Nov 24 07:53:49 crc kubenswrapper[4809]: I1124 07:53:49.745712 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zxwwc" Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.019191 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bmn4k" Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.075827 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17f65a89-36a1-48aa-b66e-c45b15bfbaf0-catalog-content\") pod \"17f65a89-36a1-48aa-b66e-c45b15bfbaf0\" (UID: \"17f65a89-36a1-48aa-b66e-c45b15bfbaf0\") " Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.075867 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17f65a89-36a1-48aa-b66e-c45b15bfbaf0-utilities\") pod \"17f65a89-36a1-48aa-b66e-c45b15bfbaf0\" (UID: \"17f65a89-36a1-48aa-b66e-c45b15bfbaf0\") " Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.075912 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5mv\" (UniqueName: \"kubernetes.io/projected/17f65a89-36a1-48aa-b66e-c45b15bfbaf0-kube-api-access-vt5mv\") pod \"17f65a89-36a1-48aa-b66e-c45b15bfbaf0\" (UID: \"17f65a89-36a1-48aa-b66e-c45b15bfbaf0\") " Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.077828 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17f65a89-36a1-48aa-b66e-c45b15bfbaf0-utilities" (OuterVolumeSpecName: "utilities") pod "17f65a89-36a1-48aa-b66e-c45b15bfbaf0" (UID: "17f65a89-36a1-48aa-b66e-c45b15bfbaf0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.087268 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17f65a89-36a1-48aa-b66e-c45b15bfbaf0-kube-api-access-vt5mv" (OuterVolumeSpecName: "kube-api-access-vt5mv") pod "17f65a89-36a1-48aa-b66e-c45b15bfbaf0" (UID: "17f65a89-36a1-48aa-b66e-c45b15bfbaf0"). InnerVolumeSpecName "kube-api-access-vt5mv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.098622 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17f65a89-36a1-48aa-b66e-c45b15bfbaf0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "17f65a89-36a1-48aa-b66e-c45b15bfbaf0" (UID: "17f65a89-36a1-48aa-b66e-c45b15bfbaf0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.178727 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17f65a89-36a1-48aa-b66e-c45b15bfbaf0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.178766 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17f65a89-36a1-48aa-b66e-c45b15bfbaf0-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.178778 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5mv\" (UniqueName: \"kubernetes.io/projected/17f65a89-36a1-48aa-b66e-c45b15bfbaf0-kube-api-access-vt5mv\") on node \"crc\" DevicePath \"\"" Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.516613 4809 generic.go:334] "Generic (PLEG): container finished" podID="17f65a89-36a1-48aa-b66e-c45b15bfbaf0" containerID="950dc8d1ea7e82bd660d42d4d4baee8493b79dccfa2400a3babadee502e0e420" exitCode=0 Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.516674 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bmn4k" Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.516708 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bmn4k" event={"ID":"17f65a89-36a1-48aa-b66e-c45b15bfbaf0","Type":"ContainerDied","Data":"950dc8d1ea7e82bd660d42d4d4baee8493b79dccfa2400a3babadee502e0e420"} Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.518452 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bmn4k" event={"ID":"17f65a89-36a1-48aa-b66e-c45b15bfbaf0","Type":"ContainerDied","Data":"213e9b0b805c24f2e449ac9d33681656d12cbd66cdcf8e6d61d1c48a819f470c"} Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.518510 4809 scope.go:117] "RemoveContainer" containerID="950dc8d1ea7e82bd660d42d4d4baee8493b79dccfa2400a3babadee502e0e420" Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.555141 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bmn4k"] Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.557401 4809 scope.go:117] "RemoveContainer" containerID="a6fcff21dec9f2ea855b7120977cbc1854a4880b1f3e8cb79fab8ce40ddf4ac8" Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.562449 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bmn4k"] Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.581430 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zxwwc" Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.581972 4809 scope.go:117] "RemoveContainer" containerID="7de3fcf2ff39acf4f061e814a3f48e4b2d911a557b5ff7cd817a8a2e0b1d80d7" Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.626680 4809 scope.go:117] "RemoveContainer" containerID="950dc8d1ea7e82bd660d42d4d4baee8493b79dccfa2400a3babadee502e0e420" Nov 24 07:53:50 crc kubenswrapper[4809]: E1124 07:53:50.627218 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"950dc8d1ea7e82bd660d42d4d4baee8493b79dccfa2400a3babadee502e0e420\": container with ID starting with 950dc8d1ea7e82bd660d42d4d4baee8493b79dccfa2400a3babadee502e0e420 not found: ID does not exist" containerID="950dc8d1ea7e82bd660d42d4d4baee8493b79dccfa2400a3babadee502e0e420" Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.627268 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"950dc8d1ea7e82bd660d42d4d4baee8493b79dccfa2400a3babadee502e0e420"} err="failed to get container status \"950dc8d1ea7e82bd660d42d4d4baee8493b79dccfa2400a3babadee502e0e420\": rpc error: code = NotFound desc = could not find container \"950dc8d1ea7e82bd660d42d4d4baee8493b79dccfa2400a3babadee502e0e420\": container with ID starting with 950dc8d1ea7e82bd660d42d4d4baee8493b79dccfa2400a3babadee502e0e420 not found: ID does not exist" Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.627303 4809 scope.go:117] "RemoveContainer" containerID="a6fcff21dec9f2ea855b7120977cbc1854a4880b1f3e8cb79fab8ce40ddf4ac8" Nov 24 07:53:50 crc kubenswrapper[4809]: E1124 07:53:50.627742 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6fcff21dec9f2ea855b7120977cbc1854a4880b1f3e8cb79fab8ce40ddf4ac8\": container with ID starting with a6fcff21dec9f2ea855b7120977cbc1854a4880b1f3e8cb79fab8ce40ddf4ac8 not found: ID does not exist" containerID="a6fcff21dec9f2ea855b7120977cbc1854a4880b1f3e8cb79fab8ce40ddf4ac8" Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.627771 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6fcff21dec9f2ea855b7120977cbc1854a4880b1f3e8cb79fab8ce40ddf4ac8"} err="failed to get container status \"a6fcff21dec9f2ea855b7120977cbc1854a4880b1f3e8cb79fab8ce40ddf4ac8\": rpc error: code = NotFound desc = could not find container \"a6fcff21dec9f2ea855b7120977cbc1854a4880b1f3e8cb79fab8ce40ddf4ac8\": container with ID starting with a6fcff21dec9f2ea855b7120977cbc1854a4880b1f3e8cb79fab8ce40ddf4ac8 not found: ID does not exist" Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.627794 4809 scope.go:117] "RemoveContainer" containerID="7de3fcf2ff39acf4f061e814a3f48e4b2d911a557b5ff7cd817a8a2e0b1d80d7" Nov 24 07:53:50 crc kubenswrapper[4809]: E1124 07:53:50.628125 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7de3fcf2ff39acf4f061e814a3f48e4b2d911a557b5ff7cd817a8a2e0b1d80d7\": container with ID starting with 7de3fcf2ff39acf4f061e814a3f48e4b2d911a557b5ff7cd817a8a2e0b1d80d7 not found: ID does not exist" containerID="7de3fcf2ff39acf4f061e814a3f48e4b2d911a557b5ff7cd817a8a2e0b1d80d7" Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.628161 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7de3fcf2ff39acf4f061e814a3f48e4b2d911a557b5ff7cd817a8a2e0b1d80d7"} err="failed to get container status \"7de3fcf2ff39acf4f061e814a3f48e4b2d911a557b5ff7cd817a8a2e0b1d80d7\": rpc error: code = NotFound desc = could not find container \"7de3fcf2ff39acf4f061e814a3f48e4b2d911a557b5ff7cd817a8a2e0b1d80d7\": container with ID starting with 7de3fcf2ff39acf4f061e814a3f48e4b2d911a557b5ff7cd817a8a2e0b1d80d7 not found: ID does not exist" Nov 24 07:53:50 crc kubenswrapper[4809]: I1124 07:53:50.904467 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17f65a89-36a1-48aa-b66e-c45b15bfbaf0" path="/var/lib/kubelet/pods/17f65a89-36a1-48aa-b66e-c45b15bfbaf0/volumes" Nov 24 07:53:52 crc kubenswrapper[4809]: I1124 07:53:52.087355 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zxwwc"] Nov 24 07:53:53 crc kubenswrapper[4809]: I1124 07:53:53.550332 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zxwwc" podUID="7dd5163d-c45a-4fb9-a369-aef70a9739e9" containerName="registry-server" containerID="cri-o://a7e7552aafe46e8ab3d4c5f8f1b25e1cbf7b2104ecd7f92c3af014709e60d7e9" gracePeriod=2 Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.048121 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zxwwc" Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.149669 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dd5163d-c45a-4fb9-a369-aef70a9739e9-catalog-content\") pod \"7dd5163d-c45a-4fb9-a369-aef70a9739e9\" (UID: \"7dd5163d-c45a-4fb9-a369-aef70a9739e9\") " Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.149725 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrl4j\" (UniqueName: \"kubernetes.io/projected/7dd5163d-c45a-4fb9-a369-aef70a9739e9-kube-api-access-lrl4j\") pod \"7dd5163d-c45a-4fb9-a369-aef70a9739e9\" (UID: \"7dd5163d-c45a-4fb9-a369-aef70a9739e9\") " Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.149921 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dd5163d-c45a-4fb9-a369-aef70a9739e9-utilities\") pod \"7dd5163d-c45a-4fb9-a369-aef70a9739e9\" (UID: \"7dd5163d-c45a-4fb9-a369-aef70a9739e9\") " Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.150650 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dd5163d-c45a-4fb9-a369-aef70a9739e9-utilities" (OuterVolumeSpecName: "utilities") pod "7dd5163d-c45a-4fb9-a369-aef70a9739e9" (UID: "7dd5163d-c45a-4fb9-a369-aef70a9739e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.151116 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dd5163d-c45a-4fb9-a369-aef70a9739e9-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.158659 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dd5163d-c45a-4fb9-a369-aef70a9739e9-kube-api-access-lrl4j" (OuterVolumeSpecName: "kube-api-access-lrl4j") pod "7dd5163d-c45a-4fb9-a369-aef70a9739e9" (UID: "7dd5163d-c45a-4fb9-a369-aef70a9739e9"). InnerVolumeSpecName "kube-api-access-lrl4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.202224 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dd5163d-c45a-4fb9-a369-aef70a9739e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7dd5163d-c45a-4fb9-a369-aef70a9739e9" (UID: "7dd5163d-c45a-4fb9-a369-aef70a9739e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.252351 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dd5163d-c45a-4fb9-a369-aef70a9739e9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.252385 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrl4j\" (UniqueName: \"kubernetes.io/projected/7dd5163d-c45a-4fb9-a369-aef70a9739e9-kube-api-access-lrl4j\") on node \"crc\" DevicePath \"\"" Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.562360 4809 generic.go:334] "Generic (PLEG): container finished" podID="7dd5163d-c45a-4fb9-a369-aef70a9739e9" containerID="a7e7552aafe46e8ab3d4c5f8f1b25e1cbf7b2104ecd7f92c3af014709e60d7e9" exitCode=0 Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.562399 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxwwc" event={"ID":"7dd5163d-c45a-4fb9-a369-aef70a9739e9","Type":"ContainerDied","Data":"a7e7552aafe46e8ab3d4c5f8f1b25e1cbf7b2104ecd7f92c3af014709e60d7e9"} Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.562424 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxwwc" event={"ID":"7dd5163d-c45a-4fb9-a369-aef70a9739e9","Type":"ContainerDied","Data":"1c84d19ddcd6e41c0ce4dc4d629b634739a08ff4b03e1fce6ab001d60d65d1a8"} Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.562441 4809 scope.go:117] "RemoveContainer" containerID="a7e7552aafe46e8ab3d4c5f8f1b25e1cbf7b2104ecd7f92c3af014709e60d7e9" Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.562453 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zxwwc" Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.594865 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zxwwc"] Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.605421 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zxwwc"] Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.608429 4809 scope.go:117] "RemoveContainer" containerID="a8bcdf784693d98e9d26622ae75814814ded29c657a773b229a155f52ca7caea" Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.629830 4809 scope.go:117] "RemoveContainer" containerID="7862ee532a8042ac7e5cd0e2f0bcfab5881ff10710a32b373ece5a546f3778b6" Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.670532 4809 scope.go:117] "RemoveContainer" containerID="a7e7552aafe46e8ab3d4c5f8f1b25e1cbf7b2104ecd7f92c3af014709e60d7e9" Nov 24 07:53:54 crc kubenswrapper[4809]: E1124 07:53:54.672316 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7e7552aafe46e8ab3d4c5f8f1b25e1cbf7b2104ecd7f92c3af014709e60d7e9\": container with ID starting with a7e7552aafe46e8ab3d4c5f8f1b25e1cbf7b2104ecd7f92c3af014709e60d7e9 not found: ID does not exist" containerID="a7e7552aafe46e8ab3d4c5f8f1b25e1cbf7b2104ecd7f92c3af014709e60d7e9" Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.672433 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7e7552aafe46e8ab3d4c5f8f1b25e1cbf7b2104ecd7f92c3af014709e60d7e9"} err="failed to get container status \"a7e7552aafe46e8ab3d4c5f8f1b25e1cbf7b2104ecd7f92c3af014709e60d7e9\": rpc error: code = NotFound desc = could not find container \"a7e7552aafe46e8ab3d4c5f8f1b25e1cbf7b2104ecd7f92c3af014709e60d7e9\": container with ID starting with a7e7552aafe46e8ab3d4c5f8f1b25e1cbf7b2104ecd7f92c3af014709e60d7e9 not found: ID does not exist" Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.672518 4809 scope.go:117] "RemoveContainer" containerID="a8bcdf784693d98e9d26622ae75814814ded29c657a773b229a155f52ca7caea" Nov 24 07:53:54 crc kubenswrapper[4809]: E1124 07:53:54.673068 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8bcdf784693d98e9d26622ae75814814ded29c657a773b229a155f52ca7caea\": container with ID starting with a8bcdf784693d98e9d26622ae75814814ded29c657a773b229a155f52ca7caea not found: ID does not exist" containerID="a8bcdf784693d98e9d26622ae75814814ded29c657a773b229a155f52ca7caea" Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.673123 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8bcdf784693d98e9d26622ae75814814ded29c657a773b229a155f52ca7caea"} err="failed to get container status \"a8bcdf784693d98e9d26622ae75814814ded29c657a773b229a155f52ca7caea\": rpc error: code = NotFound desc = could not find container \"a8bcdf784693d98e9d26622ae75814814ded29c657a773b229a155f52ca7caea\": container with ID starting with a8bcdf784693d98e9d26622ae75814814ded29c657a773b229a155f52ca7caea not found: ID does not exist" Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.673158 4809 scope.go:117] "RemoveContainer" containerID="7862ee532a8042ac7e5cd0e2f0bcfab5881ff10710a32b373ece5a546f3778b6" Nov 24 07:53:54 crc kubenswrapper[4809]: E1124 07:53:54.673795 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7862ee532a8042ac7e5cd0e2f0bcfab5881ff10710a32b373ece5a546f3778b6\": container with ID starting with 7862ee532a8042ac7e5cd0e2f0bcfab5881ff10710a32b373ece5a546f3778b6 not found: ID does not exist" containerID="7862ee532a8042ac7e5cd0e2f0bcfab5881ff10710a32b373ece5a546f3778b6" Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.673909 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7862ee532a8042ac7e5cd0e2f0bcfab5881ff10710a32b373ece5a546f3778b6"} err="failed to get container status \"7862ee532a8042ac7e5cd0e2f0bcfab5881ff10710a32b373ece5a546f3778b6\": rpc error: code = NotFound desc = could not find container \"7862ee532a8042ac7e5cd0e2f0bcfab5881ff10710a32b373ece5a546f3778b6\": container with ID starting with 7862ee532a8042ac7e5cd0e2f0bcfab5881ff10710a32b373ece5a546f3778b6 not found: ID does not exist" Nov 24 07:53:54 crc kubenswrapper[4809]: I1124 07:53:54.904390 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dd5163d-c45a-4fb9-a369-aef70a9739e9" path="/var/lib/kubelet/pods/7dd5163d-c45a-4fb9-a369-aef70a9739e9/volumes" Nov 24 07:53:57 crc kubenswrapper[4809]: I1124 07:53:57.965885 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-8p4jp_00f3a099-57a2-4da1-bb22-12c48c359ab7/kube-rbac-proxy/0.log" Nov 24 07:53:58 crc kubenswrapper[4809]: I1124 07:53:58.101224 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-8p4jp_00f3a099-57a2-4da1-bb22-12c48c359ab7/controller/0.log" Nov 24 07:53:58 crc kubenswrapper[4809]: I1124 07:53:58.188618 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-b922n_ab5d28f0-1848-47f6-bdc8-444e258af818/frr-k8s-webhook-server/0.log" Nov 24 07:53:58 crc kubenswrapper[4809]: I1124 07:53:58.329091 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-frr-files/0.log" Nov 24 07:53:58 crc kubenswrapper[4809]: I1124 07:53:58.487563 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-reloader/0.log" Nov 24 07:53:58 crc kubenswrapper[4809]: I1124 07:53:58.538262 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-metrics/0.log" Nov 24 07:53:58 crc kubenswrapper[4809]: I1124 07:53:58.545801 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-frr-files/0.log" Nov 24 07:53:58 crc kubenswrapper[4809]: I1124 07:53:58.569676 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-reloader/0.log" Nov 24 07:53:58 crc kubenswrapper[4809]: I1124 07:53:58.832379 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-metrics/0.log" Nov 24 07:53:58 crc kubenswrapper[4809]: I1124 07:53:58.844782 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-reloader/0.log" Nov 24 07:53:58 crc kubenswrapper[4809]: I1124 07:53:58.885915 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-metrics/0.log" Nov 24 07:53:58 crc kubenswrapper[4809]: I1124 07:53:58.906350 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-frr-files/0.log" Nov 24 07:53:59 crc kubenswrapper[4809]: I1124 07:53:59.097840 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-frr-files/0.log" Nov 24 07:53:59 crc kubenswrapper[4809]: I1124 07:53:59.118401 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-reloader/0.log" Nov 24 07:53:59 crc kubenswrapper[4809]: I1124 07:53:59.133087 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/controller/0.log" Nov 24 07:53:59 crc kubenswrapper[4809]: I1124 07:53:59.140331 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-metrics/0.log" Nov 24 07:53:59 crc kubenswrapper[4809]: I1124 07:53:59.312415 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/frr-metrics/0.log" Nov 24 07:53:59 crc kubenswrapper[4809]: I1124 07:53:59.380638 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/kube-rbac-proxy-frr/0.log" Nov 24 07:53:59 crc kubenswrapper[4809]: I1124 07:53:59.381099 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/kube-rbac-proxy/0.log" Nov 24 07:53:59 crc kubenswrapper[4809]: I1124 07:53:59.562562 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/reloader/0.log" Nov 24 07:53:59 crc kubenswrapper[4809]: I1124 07:53:59.638479 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7d8556447d-spn4w_a9a36b6c-e79c-4180-9118-146f3d280f15/manager/0.log" Nov 24 07:53:59 crc kubenswrapper[4809]: I1124 07:53:59.852090 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5969b4cdf8-zmvtt_baa671e1-bc87-45e6-8c04-33f408289fb4/webhook-server/0.log" Nov 24 07:54:00 crc kubenswrapper[4809]: I1124 07:54:00.079418 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-fpksh_13297f30-8d99-42d0-9760-9f88705b691f/kube-rbac-proxy/0.log" Nov 24 07:54:00 crc kubenswrapper[4809]: I1124 07:54:00.634138 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-fpksh_13297f30-8d99-42d0-9760-9f88705b691f/speaker/0.log" Nov 24 07:54:00 crc kubenswrapper[4809]: I1124 07:54:00.866923 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/frr/0.log" Nov 24 07:54:12 crc kubenswrapper[4809]: I1124 07:54:12.652303 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv_4bac6145-551f-4994-83ea-86181d6719ae/util/0.log" Nov 24 07:54:12 crc kubenswrapper[4809]: I1124 07:54:12.870492 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv_4bac6145-551f-4994-83ea-86181d6719ae/pull/0.log" Nov 24 07:54:12 crc kubenswrapper[4809]: I1124 07:54:12.884880 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv_4bac6145-551f-4994-83ea-86181d6719ae/util/0.log" Nov 24 07:54:12 crc kubenswrapper[4809]: I1124 07:54:12.941543 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv_4bac6145-551f-4994-83ea-86181d6719ae/pull/0.log" Nov 24 07:54:13 crc kubenswrapper[4809]: I1124 07:54:13.107660 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv_4bac6145-551f-4994-83ea-86181d6719ae/extract/0.log" Nov 24 07:54:13 crc kubenswrapper[4809]: I1124 07:54:13.135294 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv_4bac6145-551f-4994-83ea-86181d6719ae/util/0.log" Nov 24 07:54:13 crc kubenswrapper[4809]: I1124 07:54:13.137582 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv_4bac6145-551f-4994-83ea-86181d6719ae/pull/0.log" Nov 24 07:54:13 crc kubenswrapper[4809]: I1124 07:54:13.280960 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7lv57_a56840cb-71cf-40df-8df8-c91866948a88/extract-utilities/0.log" Nov 24 07:54:13 crc kubenswrapper[4809]: I1124 07:54:13.439587 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7lv57_a56840cb-71cf-40df-8df8-c91866948a88/extract-utilities/0.log" Nov 24 07:54:13 crc kubenswrapper[4809]: I1124 07:54:13.454300 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7lv57_a56840cb-71cf-40df-8df8-c91866948a88/extract-content/0.log" Nov 24 07:54:13 crc kubenswrapper[4809]: I1124 07:54:13.474050 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7lv57_a56840cb-71cf-40df-8df8-c91866948a88/extract-content/0.log" Nov 24 07:54:13 crc kubenswrapper[4809]: I1124 07:54:13.658300 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7lv57_a56840cb-71cf-40df-8df8-c91866948a88/extract-content/0.log" Nov 24 07:54:13 crc kubenswrapper[4809]: I1124 07:54:13.669193 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7lv57_a56840cb-71cf-40df-8df8-c91866948a88/extract-utilities/0.log" Nov 24 07:54:13 crc kubenswrapper[4809]: I1124 07:54:13.975481 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ws6wx_422d766e-1186-4b50-b09c-2fcced744b5d/extract-utilities/0.log" Nov 24 07:54:14 crc kubenswrapper[4809]: I1124 07:54:14.175513 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ws6wx_422d766e-1186-4b50-b09c-2fcced744b5d/extract-content/0.log" Nov 24 07:54:14 crc kubenswrapper[4809]: I1124 07:54:14.259204 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7lv57_a56840cb-71cf-40df-8df8-c91866948a88/registry-server/0.log" Nov 24 07:54:14 crc kubenswrapper[4809]: I1124 07:54:14.317843 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ws6wx_422d766e-1186-4b50-b09c-2fcced744b5d/extract-content/0.log" Nov 24 07:54:14 crc kubenswrapper[4809]: I1124 07:54:14.321055 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ws6wx_422d766e-1186-4b50-b09c-2fcced744b5d/extract-utilities/0.log" Nov 24 07:54:14 crc kubenswrapper[4809]: I1124 07:54:14.389736 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ws6wx_422d766e-1186-4b50-b09c-2fcced744b5d/extract-utilities/0.log" Nov 24 07:54:14 crc kubenswrapper[4809]: I1124 07:54:14.473339 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ws6wx_422d766e-1186-4b50-b09c-2fcced744b5d/extract-content/0.log" Nov 24 07:54:14 crc kubenswrapper[4809]: I1124 07:54:14.588678 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8_7feb42ed-e415-44b8-a837-da74daf4aa52/util/0.log" Nov 24 07:54:14 crc kubenswrapper[4809]: I1124 07:54:14.779629 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ws6wx_422d766e-1186-4b50-b09c-2fcced744b5d/registry-server/0.log" Nov 24 07:54:14 crc kubenswrapper[4809]: I1124 07:54:14.827232 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8_7feb42ed-e415-44b8-a837-da74daf4aa52/pull/0.log" Nov 24 07:54:14 crc kubenswrapper[4809]: I1124 07:54:14.855930 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8_7feb42ed-e415-44b8-a837-da74daf4aa52/util/0.log" Nov 24 07:54:14 crc kubenswrapper[4809]: I1124 07:54:14.908610 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8_7feb42ed-e415-44b8-a837-da74daf4aa52/pull/0.log" Nov 24 07:54:15 crc kubenswrapper[4809]: I1124 07:54:15.024152 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8_7feb42ed-e415-44b8-a837-da74daf4aa52/extract/0.log" Nov 24 07:54:15 crc kubenswrapper[4809]: I1124 07:54:15.035619 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8_7feb42ed-e415-44b8-a837-da74daf4aa52/pull/0.log" Nov 24 07:54:15 crc kubenswrapper[4809]: I1124 07:54:15.085369 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8_7feb42ed-e415-44b8-a837-da74daf4aa52/util/0.log" Nov 24 07:54:15 crc kubenswrapper[4809]: I1124 07:54:15.214185 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-zqwx5_4b7632a5-d284-4a30-89dc-c301a7d13609/marketplace-operator/0.log" Nov 24 07:54:15 crc kubenswrapper[4809]: I1124 07:54:15.241094 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-l9v7r_9586cac6-9d7e-4c04-9400-b0760540e65c/extract-utilities/0.log" Nov 24 07:54:15 crc kubenswrapper[4809]: I1124 07:54:15.486660 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-l9v7r_9586cac6-9d7e-4c04-9400-b0760540e65c/extract-content/0.log" Nov 24 07:54:15 crc kubenswrapper[4809]: I1124 07:54:15.509357 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-l9v7r_9586cac6-9d7e-4c04-9400-b0760540e65c/extract-utilities/0.log" Nov 24 07:54:15 crc kubenswrapper[4809]: I1124 07:54:15.523568 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-l9v7r_9586cac6-9d7e-4c04-9400-b0760540e65c/extract-content/0.log" Nov 24 07:54:15 crc kubenswrapper[4809]: I1124 07:54:15.670477 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-l9v7r_9586cac6-9d7e-4c04-9400-b0760540e65c/extract-content/0.log" Nov 24 07:54:15 crc kubenswrapper[4809]: I1124 07:54:15.721685 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-l9v7r_9586cac6-9d7e-4c04-9400-b0760540e65c/extract-utilities/0.log" Nov 24 07:54:15 crc kubenswrapper[4809]: I1124 07:54:15.822000 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-l9v7r_9586cac6-9d7e-4c04-9400-b0760540e65c/registry-server/0.log" Nov 24 07:54:15 crc kubenswrapper[4809]: I1124 07:54:15.886784 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nzgcn_17720a8a-d721-431b-8bc4-1efd5ac1a634/extract-utilities/0.log" Nov 24 07:54:16 crc kubenswrapper[4809]: I1124 07:54:16.081702 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nzgcn_17720a8a-d721-431b-8bc4-1efd5ac1a634/extract-utilities/0.log" Nov 24 07:54:16 crc kubenswrapper[4809]: I1124 07:54:16.127836 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nzgcn_17720a8a-d721-431b-8bc4-1efd5ac1a634/extract-content/0.log" Nov 24 07:54:16 crc kubenswrapper[4809]: I1124 07:54:16.138152 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nzgcn_17720a8a-d721-431b-8bc4-1efd5ac1a634/extract-content/0.log" Nov 24 07:54:16 crc kubenswrapper[4809]: I1124 07:54:16.290774 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nzgcn_17720a8a-d721-431b-8bc4-1efd5ac1a634/extract-utilities/0.log" Nov 24 07:54:16 crc kubenswrapper[4809]: I1124 07:54:16.302986 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nzgcn_17720a8a-d721-431b-8bc4-1efd5ac1a634/extract-content/0.log" Nov 24 07:54:16 crc kubenswrapper[4809]: I1124 07:54:16.430804 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nzgcn_17720a8a-d721-431b-8bc4-1efd5ac1a634/registry-server/0.log" Nov 24 07:54:18 crc kubenswrapper[4809]: I1124 07:54:18.043104 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:54:18 crc kubenswrapper[4809]: I1124 07:54:18.043165 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:54:48 crc kubenswrapper[4809]: I1124 07:54:48.043151 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:54:48 crc kubenswrapper[4809]: I1124 07:54:48.043814 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:54:48 crc kubenswrapper[4809]: I1124 07:54:48.043883 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 07:54:48 crc kubenswrapper[4809]: I1124 07:54:48.044852 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0"} pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:54:48 crc kubenswrapper[4809]: I1124 07:54:48.044909 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" containerID="cri-o://2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" gracePeriod=600 Nov 24 07:54:48 crc kubenswrapper[4809]: E1124 07:54:48.169595 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:54:49 crc kubenswrapper[4809]: I1124 07:54:49.075860 4809 generic.go:334] "Generic (PLEG): container finished" podID="be8050d5-1aea-453e-a06e-87702b0e856b" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" exitCode=0 Nov 24 07:54:49 crc kubenswrapper[4809]: I1124 07:54:49.076085 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerDied","Data":"2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0"} Nov 24 07:54:49 crc kubenswrapper[4809]: I1124 07:54:49.077201 4809 scope.go:117] "RemoveContainer" containerID="7d5c0ef1309a4cda8c3a374bbff14973c66d06a10b1ee334b7a60d88f824f490" Nov 24 07:54:49 crc kubenswrapper[4809]: I1124 07:54:49.077921 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:54:49 crc kubenswrapper[4809]: E1124 07:54:49.078414 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:55:00 crc kubenswrapper[4809]: I1124 07:55:00.911794 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:55:00 crc kubenswrapper[4809]: E1124 07:55:00.913364 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:55:11 crc kubenswrapper[4809]: I1124 07:55:11.893050 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:55:11 crc kubenswrapper[4809]: E1124 07:55:11.895042 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:55:22 crc kubenswrapper[4809]: I1124 07:55:22.891734 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:55:22 crc kubenswrapper[4809]: E1124 07:55:22.892786 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:55:33 crc kubenswrapper[4809]: I1124 07:55:33.894340 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:55:33 crc kubenswrapper[4809]: E1124 07:55:33.895403 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:55:48 crc kubenswrapper[4809]: I1124 07:55:48.891589 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:55:48 crc kubenswrapper[4809]: E1124 07:55:48.892370 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:55:55 crc kubenswrapper[4809]: I1124 07:55:55.131188 4809 generic.go:334] "Generic (PLEG): container finished" podID="563bc0df-08a9-42e8-a830-89c568611d6a" containerID="3464206eae56260f7d90a77cc37ab4243f12fd71622139823ed2e1c237bcde0c" exitCode=0 Nov 24 07:55:55 crc kubenswrapper[4809]: I1124 07:55:55.131310 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2bscz/must-gather-d5cpr" event={"ID":"563bc0df-08a9-42e8-a830-89c568611d6a","Type":"ContainerDied","Data":"3464206eae56260f7d90a77cc37ab4243f12fd71622139823ed2e1c237bcde0c"} Nov 24 07:55:55 crc kubenswrapper[4809]: I1124 07:55:55.132877 4809 scope.go:117] "RemoveContainer" containerID="3464206eae56260f7d90a77cc37ab4243f12fd71622139823ed2e1c237bcde0c" Nov 24 07:55:55 crc kubenswrapper[4809]: I1124 07:55:55.653444 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2bscz_must-gather-d5cpr_563bc0df-08a9-42e8-a830-89c568611d6a/gather/0.log" Nov 24 07:56:03 crc kubenswrapper[4809]: I1124 07:56:03.522776 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2bscz/must-gather-d5cpr"] Nov 24 07:56:03 crc kubenswrapper[4809]: I1124 07:56:03.523514 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-2bscz/must-gather-d5cpr" podUID="563bc0df-08a9-42e8-a830-89c568611d6a" containerName="copy" containerID="cri-o://efc18cc66fb1ea553b0d7463960298ad7dbf6941c29f3160f1bf34af34e8a12c" gracePeriod=2 Nov 24 07:56:03 crc kubenswrapper[4809]: I1124 07:56:03.535162 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2bscz/must-gather-d5cpr"] Nov 24 07:56:03 crc kubenswrapper[4809]: I1124 07:56:03.890926 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:56:03 crc kubenswrapper[4809]: E1124 07:56:03.891414 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:56:03 crc kubenswrapper[4809]: I1124 07:56:03.964474 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2bscz_must-gather-d5cpr_563bc0df-08a9-42e8-a830-89c568611d6a/copy/0.log" Nov 24 07:56:03 crc kubenswrapper[4809]: I1124 07:56:03.965528 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2bscz/must-gather-d5cpr" Nov 24 07:56:04 crc kubenswrapper[4809]: I1124 07:56:04.007939 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9clx\" (UniqueName: \"kubernetes.io/projected/563bc0df-08a9-42e8-a830-89c568611d6a-kube-api-access-f9clx\") pod \"563bc0df-08a9-42e8-a830-89c568611d6a\" (UID: \"563bc0df-08a9-42e8-a830-89c568611d6a\") " Nov 24 07:56:04 crc kubenswrapper[4809]: I1124 07:56:04.008140 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/563bc0df-08a9-42e8-a830-89c568611d6a-must-gather-output\") pod \"563bc0df-08a9-42e8-a830-89c568611d6a\" (UID: \"563bc0df-08a9-42e8-a830-89c568611d6a\") " Nov 24 07:56:04 crc kubenswrapper[4809]: I1124 07:56:04.014490 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/563bc0df-08a9-42e8-a830-89c568611d6a-kube-api-access-f9clx" (OuterVolumeSpecName: "kube-api-access-f9clx") pod "563bc0df-08a9-42e8-a830-89c568611d6a" (UID: "563bc0df-08a9-42e8-a830-89c568611d6a"). InnerVolumeSpecName "kube-api-access-f9clx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:56:04 crc kubenswrapper[4809]: I1124 07:56:04.110887 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9clx\" (UniqueName: \"kubernetes.io/projected/563bc0df-08a9-42e8-a830-89c568611d6a-kube-api-access-f9clx\") on node \"crc\" DevicePath \"\"" Nov 24 07:56:04 crc kubenswrapper[4809]: I1124 07:56:04.172818 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/563bc0df-08a9-42e8-a830-89c568611d6a-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "563bc0df-08a9-42e8-a830-89c568611d6a" (UID: "563bc0df-08a9-42e8-a830-89c568611d6a"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:56:04 crc kubenswrapper[4809]: I1124 07:56:04.212915 4809 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/563bc0df-08a9-42e8-a830-89c568611d6a-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 24 07:56:04 crc kubenswrapper[4809]: I1124 07:56:04.215397 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2bscz_must-gather-d5cpr_563bc0df-08a9-42e8-a830-89c568611d6a/copy/0.log" Nov 24 07:56:04 crc kubenswrapper[4809]: I1124 07:56:04.216147 4809 generic.go:334] "Generic (PLEG): container finished" podID="563bc0df-08a9-42e8-a830-89c568611d6a" containerID="efc18cc66fb1ea553b0d7463960298ad7dbf6941c29f3160f1bf34af34e8a12c" exitCode=143 Nov 24 07:56:04 crc kubenswrapper[4809]: I1124 07:56:04.216210 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2bscz/must-gather-d5cpr" Nov 24 07:56:04 crc kubenswrapper[4809]: I1124 07:56:04.216272 4809 scope.go:117] "RemoveContainer" containerID="efc18cc66fb1ea553b0d7463960298ad7dbf6941c29f3160f1bf34af34e8a12c" Nov 24 07:56:04 crc kubenswrapper[4809]: I1124 07:56:04.237150 4809 scope.go:117] "RemoveContainer" containerID="3464206eae56260f7d90a77cc37ab4243f12fd71622139823ed2e1c237bcde0c" Nov 24 07:56:04 crc kubenswrapper[4809]: I1124 07:56:04.300669 4809 scope.go:117] "RemoveContainer" containerID="efc18cc66fb1ea553b0d7463960298ad7dbf6941c29f3160f1bf34af34e8a12c" Nov 24 07:56:04 crc kubenswrapper[4809]: E1124 07:56:04.301169 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efc18cc66fb1ea553b0d7463960298ad7dbf6941c29f3160f1bf34af34e8a12c\": container with ID starting with efc18cc66fb1ea553b0d7463960298ad7dbf6941c29f3160f1bf34af34e8a12c not found: ID does not exist" containerID="efc18cc66fb1ea553b0d7463960298ad7dbf6941c29f3160f1bf34af34e8a12c" Nov 24 07:56:04 crc kubenswrapper[4809]: I1124 07:56:04.301227 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efc18cc66fb1ea553b0d7463960298ad7dbf6941c29f3160f1bf34af34e8a12c"} err="failed to get container status \"efc18cc66fb1ea553b0d7463960298ad7dbf6941c29f3160f1bf34af34e8a12c\": rpc error: code = NotFound desc = could not find container \"efc18cc66fb1ea553b0d7463960298ad7dbf6941c29f3160f1bf34af34e8a12c\": container with ID starting with efc18cc66fb1ea553b0d7463960298ad7dbf6941c29f3160f1bf34af34e8a12c not found: ID does not exist" Nov 24 07:56:04 crc kubenswrapper[4809]: I1124 07:56:04.301257 4809 scope.go:117] "RemoveContainer" containerID="3464206eae56260f7d90a77cc37ab4243f12fd71622139823ed2e1c237bcde0c" Nov 24 07:56:04 crc kubenswrapper[4809]: E1124 07:56:04.301696 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3464206eae56260f7d90a77cc37ab4243f12fd71622139823ed2e1c237bcde0c\": container with ID starting with 3464206eae56260f7d90a77cc37ab4243f12fd71622139823ed2e1c237bcde0c not found: ID does not exist" containerID="3464206eae56260f7d90a77cc37ab4243f12fd71622139823ed2e1c237bcde0c" Nov 24 07:56:04 crc kubenswrapper[4809]: I1124 07:56:04.301728 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3464206eae56260f7d90a77cc37ab4243f12fd71622139823ed2e1c237bcde0c"} err="failed to get container status \"3464206eae56260f7d90a77cc37ab4243f12fd71622139823ed2e1c237bcde0c\": rpc error: code = NotFound desc = could not find container \"3464206eae56260f7d90a77cc37ab4243f12fd71622139823ed2e1c237bcde0c\": container with ID starting with 3464206eae56260f7d90a77cc37ab4243f12fd71622139823ed2e1c237bcde0c not found: ID does not exist" Nov 24 07:56:04 crc kubenswrapper[4809]: I1124 07:56:04.906640 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="563bc0df-08a9-42e8-a830-89c568611d6a" path="/var/lib/kubelet/pods/563bc0df-08a9-42e8-a830-89c568611d6a/volumes" Nov 24 07:56:17 crc kubenswrapper[4809]: I1124 07:56:17.891745 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:56:17 crc kubenswrapper[4809]: E1124 07:56:17.892895 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:56:30 crc kubenswrapper[4809]: I1124 07:56:30.901378 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:56:30 crc kubenswrapper[4809]: E1124 07:56:30.903157 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:56:44 crc kubenswrapper[4809]: I1124 07:56:44.891160 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:56:44 crc kubenswrapper[4809]: E1124 07:56:44.892400 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:56:59 crc kubenswrapper[4809]: I1124 07:56:59.891590 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:56:59 crc kubenswrapper[4809]: E1124 07:56:59.892480 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:57:14 crc kubenswrapper[4809]: I1124 07:57:14.892057 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:57:14 crc kubenswrapper[4809]: E1124 07:57:14.892788 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:57:25 crc kubenswrapper[4809]: I1124 07:57:25.891401 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:57:25 crc kubenswrapper[4809]: E1124 07:57:25.893796 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:57:36 crc kubenswrapper[4809]: I1124 07:57:36.891179 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:57:36 crc kubenswrapper[4809]: E1124 07:57:36.891828 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:57:48 crc kubenswrapper[4809]: I1124 07:57:48.894009 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:57:48 crc kubenswrapper[4809]: E1124 07:57:48.895585 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:58:03 crc kubenswrapper[4809]: I1124 07:58:03.894067 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:58:03 crc kubenswrapper[4809]: E1124 07:58:03.895074 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:58:10 crc kubenswrapper[4809]: I1124 07:58:10.909761 4809 scope.go:117] "RemoveContainer" containerID="1e572b1c76177f14c9977fc4104ab2a9b1198238fac01678e94232d76bd74005" Nov 24 07:58:17 crc kubenswrapper[4809]: I1124 07:58:17.891649 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:58:17 crc kubenswrapper[4809]: E1124 07:58:17.892898 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:58:29 crc kubenswrapper[4809]: I1124 07:58:29.905382 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:58:29 crc kubenswrapper[4809]: E1124 07:58:29.906476 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.434154 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8fk5f/must-gather-kfhs5"] Nov 24 07:58:33 crc kubenswrapper[4809]: E1124 07:58:33.435016 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dd5163d-c45a-4fb9-a369-aef70a9739e9" containerName="extract-content" Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.435027 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dd5163d-c45a-4fb9-a369-aef70a9739e9" containerName="extract-content" Nov 24 07:58:33 crc kubenswrapper[4809]: E1124 07:58:33.435044 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dd5163d-c45a-4fb9-a369-aef70a9739e9" containerName="extract-utilities" Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.435051 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dd5163d-c45a-4fb9-a369-aef70a9739e9" containerName="extract-utilities" Nov 24 07:58:33 crc kubenswrapper[4809]: E1124 07:58:33.435067 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="563bc0df-08a9-42e8-a830-89c568611d6a" containerName="copy" Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.435073 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="563bc0df-08a9-42e8-a830-89c568611d6a" containerName="copy" Nov 24 07:58:33 crc kubenswrapper[4809]: E1124 07:58:33.435083 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dd5163d-c45a-4fb9-a369-aef70a9739e9" containerName="registry-server" Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.435088 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dd5163d-c45a-4fb9-a369-aef70a9739e9" containerName="registry-server" Nov 24 07:58:33 crc kubenswrapper[4809]: E1124 07:58:33.435099 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17f65a89-36a1-48aa-b66e-c45b15bfbaf0" containerName="extract-content" Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.435105 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="17f65a89-36a1-48aa-b66e-c45b15bfbaf0" containerName="extract-content" Nov 24 07:58:33 crc kubenswrapper[4809]: E1124 07:58:33.435118 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="563bc0df-08a9-42e8-a830-89c568611d6a" containerName="gather" Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.435123 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="563bc0df-08a9-42e8-a830-89c568611d6a" containerName="gather" Nov 24 07:58:33 crc kubenswrapper[4809]: E1124 07:58:33.435140 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17f65a89-36a1-48aa-b66e-c45b15bfbaf0" containerName="registry-server" Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.435146 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="17f65a89-36a1-48aa-b66e-c45b15bfbaf0" containerName="registry-server" Nov 24 07:58:33 crc kubenswrapper[4809]: E1124 07:58:33.435163 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17f65a89-36a1-48aa-b66e-c45b15bfbaf0" containerName="extract-utilities" Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.435169 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="17f65a89-36a1-48aa-b66e-c45b15bfbaf0" containerName="extract-utilities" Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.435360 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="563bc0df-08a9-42e8-a830-89c568611d6a" containerName="copy" Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.435379 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dd5163d-c45a-4fb9-a369-aef70a9739e9" containerName="registry-server" Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.435394 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="17f65a89-36a1-48aa-b66e-c45b15bfbaf0" containerName="registry-server" Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.435401 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="563bc0df-08a9-42e8-a830-89c568611d6a" containerName="gather" Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.436495 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8fk5f/must-gather-kfhs5" Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.439375 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-8fk5f"/"kube-root-ca.crt" Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.439565 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-8fk5f"/"openshift-service-ca.crt" Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.455231 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-8fk5f/must-gather-kfhs5"] Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.491978 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjvvd\" (UniqueName: \"kubernetes.io/projected/d7b9f5b4-482c-406f-9f60-b5bc186d5d85-kube-api-access-tjvvd\") pod \"must-gather-kfhs5\" (UID: \"d7b9f5b4-482c-406f-9f60-b5bc186d5d85\") " pod="openshift-must-gather-8fk5f/must-gather-kfhs5" Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.492025 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d7b9f5b4-482c-406f-9f60-b5bc186d5d85-must-gather-output\") pod \"must-gather-kfhs5\" (UID: \"d7b9f5b4-482c-406f-9f60-b5bc186d5d85\") " pod="openshift-must-gather-8fk5f/must-gather-kfhs5" Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.593175 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjvvd\" (UniqueName: \"kubernetes.io/projected/d7b9f5b4-482c-406f-9f60-b5bc186d5d85-kube-api-access-tjvvd\") pod \"must-gather-kfhs5\" (UID: \"d7b9f5b4-482c-406f-9f60-b5bc186d5d85\") " pod="openshift-must-gather-8fk5f/must-gather-kfhs5" Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.593243 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d7b9f5b4-482c-406f-9f60-b5bc186d5d85-must-gather-output\") pod \"must-gather-kfhs5\" (UID: \"d7b9f5b4-482c-406f-9f60-b5bc186d5d85\") " pod="openshift-must-gather-8fk5f/must-gather-kfhs5" Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.593772 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d7b9f5b4-482c-406f-9f60-b5bc186d5d85-must-gather-output\") pod \"must-gather-kfhs5\" (UID: \"d7b9f5b4-482c-406f-9f60-b5bc186d5d85\") " pod="openshift-must-gather-8fk5f/must-gather-kfhs5" Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.618571 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjvvd\" (UniqueName: \"kubernetes.io/projected/d7b9f5b4-482c-406f-9f60-b5bc186d5d85-kube-api-access-tjvvd\") pod \"must-gather-kfhs5\" (UID: \"d7b9f5b4-482c-406f-9f60-b5bc186d5d85\") " pod="openshift-must-gather-8fk5f/must-gather-kfhs5" Nov 24 07:58:33 crc kubenswrapper[4809]: I1124 07:58:33.797134 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8fk5f/must-gather-kfhs5" Nov 24 07:58:34 crc kubenswrapper[4809]: I1124 07:58:34.321888 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-8fk5f/must-gather-kfhs5"] Nov 24 07:58:34 crc kubenswrapper[4809]: I1124 07:58:34.701332 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8fk5f/must-gather-kfhs5" event={"ID":"d7b9f5b4-482c-406f-9f60-b5bc186d5d85","Type":"ContainerStarted","Data":"9cb24a6d45b957a626a18f5830beb1316ff0e9e77d7f41433a6d62cef50b75d3"} Nov 24 07:58:34 crc kubenswrapper[4809]: I1124 07:58:34.701678 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8fk5f/must-gather-kfhs5" event={"ID":"d7b9f5b4-482c-406f-9f60-b5bc186d5d85","Type":"ContainerStarted","Data":"41a720d5f5d9cf048626df7787b0ae5e5047b3aa1770236669b550201d3f8b2a"} Nov 24 07:58:35 crc kubenswrapper[4809]: I1124 07:58:35.712129 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8fk5f/must-gather-kfhs5" event={"ID":"d7b9f5b4-482c-406f-9f60-b5bc186d5d85","Type":"ContainerStarted","Data":"a91f3fb40b90f5af6695afbd3688541f40c0d8068e5e5d7ac455b4b8c50d9fed"} Nov 24 07:58:35 crc kubenswrapper[4809]: I1124 07:58:35.729374 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-8fk5f/must-gather-kfhs5" podStartSLOduration=2.729354183 podStartE2EDuration="2.729354183s" podCreationTimestamp="2025-11-24 07:58:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:58:35.724188523 +0000 UTC m=+3815.624780238" watchObservedRunningTime="2025-11-24 07:58:35.729354183 +0000 UTC m=+3815.629945898" Nov 24 07:58:38 crc kubenswrapper[4809]: I1124 07:58:38.100749 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8fk5f/crc-debug-ghznb"] Nov 24 07:58:38 crc kubenswrapper[4809]: I1124 07:58:38.102501 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8fk5f/crc-debug-ghznb" Nov 24 07:58:38 crc kubenswrapper[4809]: I1124 07:58:38.107874 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-8fk5f"/"default-dockercfg-t9fcg" Nov 24 07:58:38 crc kubenswrapper[4809]: I1124 07:58:38.243052 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b4ce570-b998-428e-9d77-9b5feef0d569-host\") pod \"crc-debug-ghznb\" (UID: \"4b4ce570-b998-428e-9d77-9b5feef0d569\") " pod="openshift-must-gather-8fk5f/crc-debug-ghznb" Nov 24 07:58:38 crc kubenswrapper[4809]: I1124 07:58:38.243116 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9pr4\" (UniqueName: \"kubernetes.io/projected/4b4ce570-b998-428e-9d77-9b5feef0d569-kube-api-access-l9pr4\") pod \"crc-debug-ghznb\" (UID: \"4b4ce570-b998-428e-9d77-9b5feef0d569\") " pod="openshift-must-gather-8fk5f/crc-debug-ghznb" Nov 24 07:58:38 crc kubenswrapper[4809]: I1124 07:58:38.345337 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b4ce570-b998-428e-9d77-9b5feef0d569-host\") pod \"crc-debug-ghznb\" (UID: \"4b4ce570-b998-428e-9d77-9b5feef0d569\") " pod="openshift-must-gather-8fk5f/crc-debug-ghznb" Nov 24 07:58:38 crc kubenswrapper[4809]: I1124 07:58:38.345408 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9pr4\" (UniqueName: \"kubernetes.io/projected/4b4ce570-b998-428e-9d77-9b5feef0d569-kube-api-access-l9pr4\") pod \"crc-debug-ghznb\" (UID: \"4b4ce570-b998-428e-9d77-9b5feef0d569\") " pod="openshift-must-gather-8fk5f/crc-debug-ghznb" Nov 24 07:58:38 crc kubenswrapper[4809]: I1124 07:58:38.345507 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b4ce570-b998-428e-9d77-9b5feef0d569-host\") pod \"crc-debug-ghznb\" (UID: \"4b4ce570-b998-428e-9d77-9b5feef0d569\") " pod="openshift-must-gather-8fk5f/crc-debug-ghznb" Nov 24 07:58:38 crc kubenswrapper[4809]: I1124 07:58:38.368673 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9pr4\" (UniqueName: \"kubernetes.io/projected/4b4ce570-b998-428e-9d77-9b5feef0d569-kube-api-access-l9pr4\") pod \"crc-debug-ghznb\" (UID: \"4b4ce570-b998-428e-9d77-9b5feef0d569\") " pod="openshift-must-gather-8fk5f/crc-debug-ghznb" Nov 24 07:58:38 crc kubenswrapper[4809]: I1124 07:58:38.421635 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8fk5f/crc-debug-ghznb" Nov 24 07:58:38 crc kubenswrapper[4809]: I1124 07:58:38.739353 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8fk5f/crc-debug-ghznb" event={"ID":"4b4ce570-b998-428e-9d77-9b5feef0d569","Type":"ContainerStarted","Data":"5df154371da901cc8e28f07e83f7b779996e15a753150dede8d228e04403ff4c"} Nov 24 07:58:38 crc kubenswrapper[4809]: I1124 07:58:38.739653 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8fk5f/crc-debug-ghznb" event={"ID":"4b4ce570-b998-428e-9d77-9b5feef0d569","Type":"ContainerStarted","Data":"eaaf3ca8467be0eb921d94b7a5cd401dbd5ef2dd080ab112dafbe57b65ee07d6"} Nov 24 07:58:38 crc kubenswrapper[4809]: I1124 07:58:38.767594 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-8fk5f/crc-debug-ghznb" podStartSLOduration=0.767572061 podStartE2EDuration="767.572061ms" podCreationTimestamp="2025-11-24 07:58:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:58:38.753752999 +0000 UTC m=+3818.654344694" watchObservedRunningTime="2025-11-24 07:58:38.767572061 +0000 UTC m=+3818.668163766" Nov 24 07:58:43 crc kubenswrapper[4809]: I1124 07:58:43.893730 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:58:43 crc kubenswrapper[4809]: E1124 07:58:43.894483 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:58:55 crc kubenswrapper[4809]: I1124 07:58:55.891353 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:58:55 crc kubenswrapper[4809]: E1124 07:58:55.892175 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:59:06 crc kubenswrapper[4809]: I1124 07:59:06.891558 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:59:06 crc kubenswrapper[4809]: E1124 07:59:06.894157 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:59:13 crc kubenswrapper[4809]: I1124 07:59:13.079564 4809 generic.go:334] "Generic (PLEG): container finished" podID="4b4ce570-b998-428e-9d77-9b5feef0d569" containerID="5df154371da901cc8e28f07e83f7b779996e15a753150dede8d228e04403ff4c" exitCode=0 Nov 24 07:59:13 crc kubenswrapper[4809]: I1124 07:59:13.079656 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8fk5f/crc-debug-ghznb" event={"ID":"4b4ce570-b998-428e-9d77-9b5feef0d569","Type":"ContainerDied","Data":"5df154371da901cc8e28f07e83f7b779996e15a753150dede8d228e04403ff4c"} Nov 24 07:59:14 crc kubenswrapper[4809]: I1124 07:59:14.188358 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8fk5f/crc-debug-ghznb" Nov 24 07:59:14 crc kubenswrapper[4809]: I1124 07:59:14.217083 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8fk5f/crc-debug-ghznb"] Nov 24 07:59:14 crc kubenswrapper[4809]: I1124 07:59:14.225797 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8fk5f/crc-debug-ghznb"] Nov 24 07:59:14 crc kubenswrapper[4809]: I1124 07:59:14.315145 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9pr4\" (UniqueName: \"kubernetes.io/projected/4b4ce570-b998-428e-9d77-9b5feef0d569-kube-api-access-l9pr4\") pod \"4b4ce570-b998-428e-9d77-9b5feef0d569\" (UID: \"4b4ce570-b998-428e-9d77-9b5feef0d569\") " Nov 24 07:59:14 crc kubenswrapper[4809]: I1124 07:59:14.315261 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b4ce570-b998-428e-9d77-9b5feef0d569-host\") pod \"4b4ce570-b998-428e-9d77-9b5feef0d569\" (UID: \"4b4ce570-b998-428e-9d77-9b5feef0d569\") " Nov 24 07:59:14 crc kubenswrapper[4809]: I1124 07:59:14.315366 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4b4ce570-b998-428e-9d77-9b5feef0d569-host" (OuterVolumeSpecName: "host") pod "4b4ce570-b998-428e-9d77-9b5feef0d569" (UID: "4b4ce570-b998-428e-9d77-9b5feef0d569"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:59:14 crc kubenswrapper[4809]: I1124 07:59:14.315882 4809 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b4ce570-b998-428e-9d77-9b5feef0d569-host\") on node \"crc\" DevicePath \"\"" Nov 24 07:59:14 crc kubenswrapper[4809]: I1124 07:59:14.322264 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b4ce570-b998-428e-9d77-9b5feef0d569-kube-api-access-l9pr4" (OuterVolumeSpecName: "kube-api-access-l9pr4") pod "4b4ce570-b998-428e-9d77-9b5feef0d569" (UID: "4b4ce570-b998-428e-9d77-9b5feef0d569"). InnerVolumeSpecName "kube-api-access-l9pr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:59:14 crc kubenswrapper[4809]: I1124 07:59:14.417353 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9pr4\" (UniqueName: \"kubernetes.io/projected/4b4ce570-b998-428e-9d77-9b5feef0d569-kube-api-access-l9pr4\") on node \"crc\" DevicePath \"\"" Nov 24 07:59:14 crc kubenswrapper[4809]: I1124 07:59:14.903561 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b4ce570-b998-428e-9d77-9b5feef0d569" path="/var/lib/kubelet/pods/4b4ce570-b998-428e-9d77-9b5feef0d569/volumes" Nov 24 07:59:15 crc kubenswrapper[4809]: I1124 07:59:15.102947 4809 scope.go:117] "RemoveContainer" containerID="5df154371da901cc8e28f07e83f7b779996e15a753150dede8d228e04403ff4c" Nov 24 07:59:15 crc kubenswrapper[4809]: I1124 07:59:15.103006 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8fk5f/crc-debug-ghznb" Nov 24 07:59:15 crc kubenswrapper[4809]: I1124 07:59:15.432795 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8fk5f/crc-debug-8q6p5"] Nov 24 07:59:15 crc kubenswrapper[4809]: E1124 07:59:15.433490 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b4ce570-b998-428e-9d77-9b5feef0d569" containerName="container-00" Nov 24 07:59:15 crc kubenswrapper[4809]: I1124 07:59:15.433503 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b4ce570-b998-428e-9d77-9b5feef0d569" containerName="container-00" Nov 24 07:59:15 crc kubenswrapper[4809]: I1124 07:59:15.433673 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b4ce570-b998-428e-9d77-9b5feef0d569" containerName="container-00" Nov 24 07:59:15 crc kubenswrapper[4809]: I1124 07:59:15.434306 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8fk5f/crc-debug-8q6p5" Nov 24 07:59:15 crc kubenswrapper[4809]: I1124 07:59:15.436613 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-8fk5f"/"default-dockercfg-t9fcg" Nov 24 07:59:15 crc kubenswrapper[4809]: I1124 07:59:15.538300 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d449606f-fcc0-4fe8-bef3-712eba352263-host\") pod \"crc-debug-8q6p5\" (UID: \"d449606f-fcc0-4fe8-bef3-712eba352263\") " pod="openshift-must-gather-8fk5f/crc-debug-8q6p5" Nov 24 07:59:15 crc kubenswrapper[4809]: I1124 07:59:15.538410 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zqfj\" (UniqueName: \"kubernetes.io/projected/d449606f-fcc0-4fe8-bef3-712eba352263-kube-api-access-5zqfj\") pod \"crc-debug-8q6p5\" (UID: \"d449606f-fcc0-4fe8-bef3-712eba352263\") " pod="openshift-must-gather-8fk5f/crc-debug-8q6p5" Nov 24 07:59:15 crc kubenswrapper[4809]: I1124 07:59:15.640735 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d449606f-fcc0-4fe8-bef3-712eba352263-host\") pod \"crc-debug-8q6p5\" (UID: \"d449606f-fcc0-4fe8-bef3-712eba352263\") " pod="openshift-must-gather-8fk5f/crc-debug-8q6p5" Nov 24 07:59:15 crc kubenswrapper[4809]: I1124 07:59:15.641011 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d449606f-fcc0-4fe8-bef3-712eba352263-host\") pod \"crc-debug-8q6p5\" (UID: \"d449606f-fcc0-4fe8-bef3-712eba352263\") " pod="openshift-must-gather-8fk5f/crc-debug-8q6p5" Nov 24 07:59:15 crc kubenswrapper[4809]: I1124 07:59:15.641230 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zqfj\" (UniqueName: \"kubernetes.io/projected/d449606f-fcc0-4fe8-bef3-712eba352263-kube-api-access-5zqfj\") pod \"crc-debug-8q6p5\" (UID: \"d449606f-fcc0-4fe8-bef3-712eba352263\") " pod="openshift-must-gather-8fk5f/crc-debug-8q6p5" Nov 24 07:59:15 crc kubenswrapper[4809]: I1124 07:59:15.658305 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zqfj\" (UniqueName: \"kubernetes.io/projected/d449606f-fcc0-4fe8-bef3-712eba352263-kube-api-access-5zqfj\") pod \"crc-debug-8q6p5\" (UID: \"d449606f-fcc0-4fe8-bef3-712eba352263\") " pod="openshift-must-gather-8fk5f/crc-debug-8q6p5" Nov 24 07:59:15 crc kubenswrapper[4809]: I1124 07:59:15.760176 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8fk5f/crc-debug-8q6p5" Nov 24 07:59:16 crc kubenswrapper[4809]: I1124 07:59:16.113702 4809 generic.go:334] "Generic (PLEG): container finished" podID="d449606f-fcc0-4fe8-bef3-712eba352263" containerID="49a75774ffb81d989d3503b5880ffce4b994095f8396a88aec21025c86759e11" exitCode=0 Nov 24 07:59:16 crc kubenswrapper[4809]: I1124 07:59:16.113776 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8fk5f/crc-debug-8q6p5" event={"ID":"d449606f-fcc0-4fe8-bef3-712eba352263","Type":"ContainerDied","Data":"49a75774ffb81d989d3503b5880ffce4b994095f8396a88aec21025c86759e11"} Nov 24 07:59:16 crc kubenswrapper[4809]: I1124 07:59:16.113802 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8fk5f/crc-debug-8q6p5" event={"ID":"d449606f-fcc0-4fe8-bef3-712eba352263","Type":"ContainerStarted","Data":"874d48fb5eb5b677eabe57be98fcf861d267fff0fae09de0e68fb0f25549f047"} Nov 24 07:59:16 crc kubenswrapper[4809]: I1124 07:59:16.596750 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8fk5f/crc-debug-8q6p5"] Nov 24 07:59:16 crc kubenswrapper[4809]: I1124 07:59:16.604840 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8fk5f/crc-debug-8q6p5"] Nov 24 07:59:17 crc kubenswrapper[4809]: I1124 07:59:17.244177 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8fk5f/crc-debug-8q6p5" Nov 24 07:59:17 crc kubenswrapper[4809]: I1124 07:59:17.375046 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zqfj\" (UniqueName: \"kubernetes.io/projected/d449606f-fcc0-4fe8-bef3-712eba352263-kube-api-access-5zqfj\") pod \"d449606f-fcc0-4fe8-bef3-712eba352263\" (UID: \"d449606f-fcc0-4fe8-bef3-712eba352263\") " Nov 24 07:59:17 crc kubenswrapper[4809]: I1124 07:59:17.376708 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d449606f-fcc0-4fe8-bef3-712eba352263-host\") pod \"d449606f-fcc0-4fe8-bef3-712eba352263\" (UID: \"d449606f-fcc0-4fe8-bef3-712eba352263\") " Nov 24 07:59:17 crc kubenswrapper[4809]: I1124 07:59:17.376829 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d449606f-fcc0-4fe8-bef3-712eba352263-host" (OuterVolumeSpecName: "host") pod "d449606f-fcc0-4fe8-bef3-712eba352263" (UID: "d449606f-fcc0-4fe8-bef3-712eba352263"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:59:17 crc kubenswrapper[4809]: I1124 07:59:17.378119 4809 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d449606f-fcc0-4fe8-bef3-712eba352263-host\") on node \"crc\" DevicePath \"\"" Nov 24 07:59:17 crc kubenswrapper[4809]: I1124 07:59:17.381211 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d449606f-fcc0-4fe8-bef3-712eba352263-kube-api-access-5zqfj" (OuterVolumeSpecName: "kube-api-access-5zqfj") pod "d449606f-fcc0-4fe8-bef3-712eba352263" (UID: "d449606f-fcc0-4fe8-bef3-712eba352263"). InnerVolumeSpecName "kube-api-access-5zqfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:59:17 crc kubenswrapper[4809]: I1124 07:59:17.479747 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zqfj\" (UniqueName: \"kubernetes.io/projected/d449606f-fcc0-4fe8-bef3-712eba352263-kube-api-access-5zqfj\") on node \"crc\" DevicePath \"\"" Nov 24 07:59:17 crc kubenswrapper[4809]: I1124 07:59:17.757866 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8fk5f/crc-debug-w6gpq"] Nov 24 07:59:17 crc kubenswrapper[4809]: E1124 07:59:17.758298 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d449606f-fcc0-4fe8-bef3-712eba352263" containerName="container-00" Nov 24 07:59:17 crc kubenswrapper[4809]: I1124 07:59:17.758313 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d449606f-fcc0-4fe8-bef3-712eba352263" containerName="container-00" Nov 24 07:59:17 crc kubenswrapper[4809]: I1124 07:59:17.758512 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d449606f-fcc0-4fe8-bef3-712eba352263" containerName="container-00" Nov 24 07:59:17 crc kubenswrapper[4809]: I1124 07:59:17.759127 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8fk5f/crc-debug-w6gpq" Nov 24 07:59:17 crc kubenswrapper[4809]: I1124 07:59:17.784392 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc8n6\" (UniqueName: \"kubernetes.io/projected/4c6307ff-83c1-44b4-badd-367d3dad9d51-kube-api-access-pc8n6\") pod \"crc-debug-w6gpq\" (UID: \"4c6307ff-83c1-44b4-badd-367d3dad9d51\") " pod="openshift-must-gather-8fk5f/crc-debug-w6gpq" Nov 24 07:59:17 crc kubenswrapper[4809]: I1124 07:59:17.784477 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c6307ff-83c1-44b4-badd-367d3dad9d51-host\") pod \"crc-debug-w6gpq\" (UID: \"4c6307ff-83c1-44b4-badd-367d3dad9d51\") " pod="openshift-must-gather-8fk5f/crc-debug-w6gpq" Nov 24 07:59:17 crc kubenswrapper[4809]: I1124 07:59:17.886762 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc8n6\" (UniqueName: \"kubernetes.io/projected/4c6307ff-83c1-44b4-badd-367d3dad9d51-kube-api-access-pc8n6\") pod \"crc-debug-w6gpq\" (UID: \"4c6307ff-83c1-44b4-badd-367d3dad9d51\") " pod="openshift-must-gather-8fk5f/crc-debug-w6gpq" Nov 24 07:59:17 crc kubenswrapper[4809]: I1124 07:59:17.886818 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c6307ff-83c1-44b4-badd-367d3dad9d51-host\") pod \"crc-debug-w6gpq\" (UID: \"4c6307ff-83c1-44b4-badd-367d3dad9d51\") " pod="openshift-must-gather-8fk5f/crc-debug-w6gpq" Nov 24 07:59:17 crc kubenswrapper[4809]: I1124 07:59:17.886972 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c6307ff-83c1-44b4-badd-367d3dad9d51-host\") pod \"crc-debug-w6gpq\" (UID: \"4c6307ff-83c1-44b4-badd-367d3dad9d51\") " pod="openshift-must-gather-8fk5f/crc-debug-w6gpq" Nov 24 07:59:17 crc kubenswrapper[4809]: I1124 07:59:17.905148 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc8n6\" (UniqueName: \"kubernetes.io/projected/4c6307ff-83c1-44b4-badd-367d3dad9d51-kube-api-access-pc8n6\") pod \"crc-debug-w6gpq\" (UID: \"4c6307ff-83c1-44b4-badd-367d3dad9d51\") " pod="openshift-must-gather-8fk5f/crc-debug-w6gpq" Nov 24 07:59:18 crc kubenswrapper[4809]: I1124 07:59:18.075772 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8fk5f/crc-debug-w6gpq" Nov 24 07:59:18 crc kubenswrapper[4809]: W1124 07:59:18.107907 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c6307ff_83c1_44b4_badd_367d3dad9d51.slice/crio-355abc77b5b74083e0746a699c9295e8be530fef6dbe91f24c59077a58968596 WatchSource:0}: Error finding container 355abc77b5b74083e0746a699c9295e8be530fef6dbe91f24c59077a58968596: Status 404 returned error can't find the container with id 355abc77b5b74083e0746a699c9295e8be530fef6dbe91f24c59077a58968596 Nov 24 07:59:18 crc kubenswrapper[4809]: I1124 07:59:18.143205 4809 scope.go:117] "RemoveContainer" containerID="49a75774ffb81d989d3503b5880ffce4b994095f8396a88aec21025c86759e11" Nov 24 07:59:18 crc kubenswrapper[4809]: I1124 07:59:18.143501 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8fk5f/crc-debug-8q6p5" Nov 24 07:59:18 crc kubenswrapper[4809]: I1124 07:59:18.148134 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8fk5f/crc-debug-w6gpq" event={"ID":"4c6307ff-83c1-44b4-badd-367d3dad9d51","Type":"ContainerStarted","Data":"355abc77b5b74083e0746a699c9295e8be530fef6dbe91f24c59077a58968596"} Nov 24 07:59:18 crc kubenswrapper[4809]: I1124 07:59:18.891950 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:59:18 crc kubenswrapper[4809]: E1124 07:59:18.892489 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:59:18 crc kubenswrapper[4809]: I1124 07:59:18.903463 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d449606f-fcc0-4fe8-bef3-712eba352263" path="/var/lib/kubelet/pods/d449606f-fcc0-4fe8-bef3-712eba352263/volumes" Nov 24 07:59:19 crc kubenswrapper[4809]: I1124 07:59:19.159684 4809 generic.go:334] "Generic (PLEG): container finished" podID="4c6307ff-83c1-44b4-badd-367d3dad9d51" containerID="df72b9caef061d59123c98e2df7c29199b7686e0907c3560ca493fb07b8ddb33" exitCode=0 Nov 24 07:59:19 crc kubenswrapper[4809]: I1124 07:59:19.159741 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8fk5f/crc-debug-w6gpq" event={"ID":"4c6307ff-83c1-44b4-badd-367d3dad9d51","Type":"ContainerDied","Data":"df72b9caef061d59123c98e2df7c29199b7686e0907c3560ca493fb07b8ddb33"} Nov 24 07:59:19 crc kubenswrapper[4809]: I1124 07:59:19.208460 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8fk5f/crc-debug-w6gpq"] Nov 24 07:59:19 crc kubenswrapper[4809]: I1124 07:59:19.214732 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8fk5f/crc-debug-w6gpq"] Nov 24 07:59:20 crc kubenswrapper[4809]: I1124 07:59:20.283855 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8fk5f/crc-debug-w6gpq" Nov 24 07:59:20 crc kubenswrapper[4809]: I1124 07:59:20.438062 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c6307ff-83c1-44b4-badd-367d3dad9d51-host\") pod \"4c6307ff-83c1-44b4-badd-367d3dad9d51\" (UID: \"4c6307ff-83c1-44b4-badd-367d3dad9d51\") " Nov 24 07:59:20 crc kubenswrapper[4809]: I1124 07:59:20.438242 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c6307ff-83c1-44b4-badd-367d3dad9d51-host" (OuterVolumeSpecName: "host") pod "4c6307ff-83c1-44b4-badd-367d3dad9d51" (UID: "4c6307ff-83c1-44b4-badd-367d3dad9d51"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:59:20 crc kubenswrapper[4809]: I1124 07:59:20.438269 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pc8n6\" (UniqueName: \"kubernetes.io/projected/4c6307ff-83c1-44b4-badd-367d3dad9d51-kube-api-access-pc8n6\") pod \"4c6307ff-83c1-44b4-badd-367d3dad9d51\" (UID: \"4c6307ff-83c1-44b4-badd-367d3dad9d51\") " Nov 24 07:59:20 crc kubenswrapper[4809]: I1124 07:59:20.439150 4809 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c6307ff-83c1-44b4-badd-367d3dad9d51-host\") on node \"crc\" DevicePath \"\"" Nov 24 07:59:20 crc kubenswrapper[4809]: I1124 07:59:20.449180 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c6307ff-83c1-44b4-badd-367d3dad9d51-kube-api-access-pc8n6" (OuterVolumeSpecName: "kube-api-access-pc8n6") pod "4c6307ff-83c1-44b4-badd-367d3dad9d51" (UID: "4c6307ff-83c1-44b4-badd-367d3dad9d51"). InnerVolumeSpecName "kube-api-access-pc8n6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:59:20 crc kubenswrapper[4809]: I1124 07:59:20.540222 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pc8n6\" (UniqueName: \"kubernetes.io/projected/4c6307ff-83c1-44b4-badd-367d3dad9d51-kube-api-access-pc8n6\") on node \"crc\" DevicePath \"\"" Nov 24 07:59:20 crc kubenswrapper[4809]: I1124 07:59:20.901533 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c6307ff-83c1-44b4-badd-367d3dad9d51" path="/var/lib/kubelet/pods/4c6307ff-83c1-44b4-badd-367d3dad9d51/volumes" Nov 24 07:59:21 crc kubenswrapper[4809]: I1124 07:59:21.179102 4809 scope.go:117] "RemoveContainer" containerID="df72b9caef061d59123c98e2df7c29199b7686e0907c3560ca493fb07b8ddb33" Nov 24 07:59:21 crc kubenswrapper[4809]: I1124 07:59:21.179142 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8fk5f/crc-debug-w6gpq" Nov 24 07:59:31 crc kubenswrapper[4809]: I1124 07:59:31.891065 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:59:31 crc kubenswrapper[4809]: E1124 07:59:31.891770 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:59:40 crc kubenswrapper[4809]: I1124 07:59:40.792054 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7bb85696db-hj9dc_04780f37-736e-4f56-a887-e50c11e3b3ac/barbican-api/0.log" Nov 24 07:59:40 crc kubenswrapper[4809]: I1124 07:59:40.894447 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7bb85696db-hj9dc_04780f37-736e-4f56-a887-e50c11e3b3ac/barbican-api-log/0.log" Nov 24 07:59:41 crc kubenswrapper[4809]: I1124 07:59:41.001289 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7947c766c8-w6s92_ada052fc-743e-477f-b5be-848ca3536f09/barbican-keystone-listener/0.log" Nov 24 07:59:41 crc kubenswrapper[4809]: I1124 07:59:41.106121 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7947c766c8-w6s92_ada052fc-743e-477f-b5be-848ca3536f09/barbican-keystone-listener-log/0.log" Nov 24 07:59:41 crc kubenswrapper[4809]: I1124 07:59:41.180838 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-cc9bfd6b9-27jr8_5c05e3fc-9407-4bc4-86b4-409ebee7f3e7/barbican-worker/0.log" Nov 24 07:59:41 crc kubenswrapper[4809]: I1124 07:59:41.206681 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-cc9bfd6b9-27jr8_5c05e3fc-9407-4bc4-86b4-409ebee7f3e7/barbican-worker-log/0.log" Nov 24 07:59:41 crc kubenswrapper[4809]: I1124 07:59:41.363935 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-rtt28_bad47712-c44f-4799-bfe2-6c95ea3b513f/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:59:41 crc kubenswrapper[4809]: I1124 07:59:41.413041 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_96597dff-85f6-4758-a2e5-27c5b01bf34a/ceilometer-central-agent/0.log" Nov 24 07:59:41 crc kubenswrapper[4809]: I1124 07:59:41.517360 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_96597dff-85f6-4758-a2e5-27c5b01bf34a/proxy-httpd/0.log" Nov 24 07:59:41 crc kubenswrapper[4809]: I1124 07:59:41.550641 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_96597dff-85f6-4758-a2e5-27c5b01bf34a/ceilometer-notification-agent/0.log" Nov 24 07:59:41 crc kubenswrapper[4809]: I1124 07:59:41.578707 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_96597dff-85f6-4758-a2e5-27c5b01bf34a/sg-core/0.log" Nov 24 07:59:41 crc kubenswrapper[4809]: I1124 07:59:41.744177 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_097f9d2f-9e7a-42b8-a40e-da15a39b146c/cinder-api/0.log" Nov 24 07:59:41 crc kubenswrapper[4809]: I1124 07:59:41.773680 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_097f9d2f-9e7a-42b8-a40e-da15a39b146c/cinder-api-log/0.log" Nov 24 07:59:41 crc kubenswrapper[4809]: I1124 07:59:41.895196 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8ea4784f-e104-4ad0-873e-7d336cf4d98e/cinder-scheduler/0.log" Nov 24 07:59:41 crc kubenswrapper[4809]: I1124 07:59:41.945403 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8ea4784f-e104-4ad0-873e-7d336cf4d98e/probe/0.log" Nov 24 07:59:42 crc kubenswrapper[4809]: I1124 07:59:42.028438 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-6qg8k_8a942ac4-7652-45aa-b238-3aff50848195/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:59:42 crc kubenswrapper[4809]: I1124 07:59:42.132959 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-hmbvr_a7f9b89a-871e-4478-bb01-1120cd72a2b6/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:59:42 crc kubenswrapper[4809]: I1124 07:59:42.241816 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-f6c2x_09eb3227-9b7b-4d9a-9fa1-80e28696c61e/init/0.log" Nov 24 07:59:42 crc kubenswrapper[4809]: I1124 07:59:42.378803 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-f6c2x_09eb3227-9b7b-4d9a-9fa1-80e28696c61e/init/0.log" Nov 24 07:59:42 crc kubenswrapper[4809]: I1124 07:59:42.426046 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-f6c2x_09eb3227-9b7b-4d9a-9fa1-80e28696c61e/dnsmasq-dns/0.log" Nov 24 07:59:42 crc kubenswrapper[4809]: I1124 07:59:42.437285 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-lwmd9_a022199a-9d09-421e-b9f3-67d00cab4f5f/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:59:42 crc kubenswrapper[4809]: I1124 07:59:42.597791 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_cfb0ea66-5650-4f85-a36a-06b60949a4bf/glance-log/0.log" Nov 24 07:59:42 crc kubenswrapper[4809]: I1124 07:59:42.604983 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_cfb0ea66-5650-4f85-a36a-06b60949a4bf/glance-httpd/0.log" Nov 24 07:59:42 crc kubenswrapper[4809]: I1124 07:59:42.782461 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_91d79979-625d-4ad3-bbb5-f3f35e7142a4/glance-log/0.log" Nov 24 07:59:42 crc kubenswrapper[4809]: I1124 07:59:42.826749 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_91d79979-625d-4ad3-bbb5-f3f35e7142a4/glance-httpd/0.log" Nov 24 07:59:42 crc kubenswrapper[4809]: I1124 07:59:42.892651 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:59:42 crc kubenswrapper[4809]: E1124 07:59:42.892852 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dr8hv_openshift-machine-config-operator(be8050d5-1aea-453e-a06e-87702b0e856b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" Nov 24 07:59:43 crc kubenswrapper[4809]: I1124 07:59:43.062102 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-8585547888-qdlwv_b5487494-236a-4335-977f-11167eaa3cdc/horizon/0.log" Nov 24 07:59:43 crc kubenswrapper[4809]: I1124 07:59:43.114555 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-jwqx9_d4e896d6-2d97-431e-bfaf-f48369c0af23/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:59:43 crc kubenswrapper[4809]: I1124 07:59:43.311319 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-2fc9j_241ce393-aede-4973-adcf-93efc73bcd32/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:59:43 crc kubenswrapper[4809]: I1124 07:59:43.372693 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-8585547888-qdlwv_b5487494-236a-4335-977f-11167eaa3cdc/horizon-log/0.log" Nov 24 07:59:43 crc kubenswrapper[4809]: I1124 07:59:43.560707 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_6f65b9e2-204d-455f-b8e1-43ccec98b7f6/kube-state-metrics/0.log" Nov 24 07:59:43 crc kubenswrapper[4809]: I1124 07:59:43.610463 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-fd8844f9c-6jj95_1a3f2bae-ab41-461b-87bc-455f850805e6/keystone-api/0.log" Nov 24 07:59:43 crc kubenswrapper[4809]: I1124 07:59:43.800615 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-x9s4s_fd55a20b-bdd2-40bf-85d7-c0b485352b8f/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:59:44 crc kubenswrapper[4809]: I1124 07:59:44.153706 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6c5d67c78c-vvvv9_9c53f941-5f7d-489c-85eb-d07ff8d4ef6b/neutron-httpd/0.log" Nov 24 07:59:44 crc kubenswrapper[4809]: I1124 07:59:44.196929 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6c5d67c78c-vvvv9_9c53f941-5f7d-489c-85eb-d07ff8d4ef6b/neutron-api/0.log" Nov 24 07:59:44 crc kubenswrapper[4809]: I1124 07:59:44.337499 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-62h7x_aef096ab-d8a4-4b78-97be-17967897eb98/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:59:44 crc kubenswrapper[4809]: I1124 07:59:44.866465 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_6e8e43ff-8f38-4981-a78c-41934ecc1210/nova-api-log/0.log" Nov 24 07:59:45 crc kubenswrapper[4809]: I1124 07:59:45.080934 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_663c08e1-fd97-42fd-b882-557ca9d71cdf/nova-cell0-conductor-conductor/0.log" Nov 24 07:59:45 crc kubenswrapper[4809]: I1124 07:59:45.340098 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_f4318b40-eb3b-4de1-8c8f-c9e4bf5996f8/nova-cell1-conductor-conductor/0.log" Nov 24 07:59:45 crc kubenswrapper[4809]: I1124 07:59:45.368716 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_6e8e43ff-8f38-4981-a78c-41934ecc1210/nova-api-api/0.log" Nov 24 07:59:45 crc kubenswrapper[4809]: I1124 07:59:45.427915 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_c0f2a430-fd7f-419b-a07c-283144e6f0f2/nova-cell1-novncproxy-novncproxy/0.log" Nov 24 07:59:45 crc kubenswrapper[4809]: I1124 07:59:45.618217 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-wsk2q_4b9c2d10-ec36-41c7-985c-44ee750eddd0/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:59:45 crc kubenswrapper[4809]: I1124 07:59:45.731790 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_d9b4155b-3ac1-4c8d-b768-e28485799816/nova-metadata-log/0.log" Nov 24 07:59:46 crc kubenswrapper[4809]: I1124 07:59:46.089032 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_d4270af3-165e-49e5-b88e-1c07b1a3a950/nova-scheduler-scheduler/0.log" Nov 24 07:59:46 crc kubenswrapper[4809]: I1124 07:59:46.141119 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e5afba7d-dbf3-4aa6-863b-1d3a6d71af88/mysql-bootstrap/0.log" Nov 24 07:59:46 crc kubenswrapper[4809]: I1124 07:59:46.319328 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e5afba7d-dbf3-4aa6-863b-1d3a6d71af88/galera/0.log" Nov 24 07:59:46 crc kubenswrapper[4809]: I1124 07:59:46.331934 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e5afba7d-dbf3-4aa6-863b-1d3a6d71af88/mysql-bootstrap/0.log" Nov 24 07:59:46 crc kubenswrapper[4809]: I1124 07:59:46.532247 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_caea018d-04b5-4d20-8161-9a9a87f67583/mysql-bootstrap/0.log" Nov 24 07:59:46 crc kubenswrapper[4809]: I1124 07:59:46.681337 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_caea018d-04b5-4d20-8161-9a9a87f67583/mysql-bootstrap/0.log" Nov 24 07:59:46 crc kubenswrapper[4809]: I1124 07:59:46.709848 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_caea018d-04b5-4d20-8161-9a9a87f67583/galera/0.log" Nov 24 07:59:46 crc kubenswrapper[4809]: I1124 07:59:46.932502 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_45b7caaf-bebc-4bef-804b-a39246b436a0/openstackclient/0.log" Nov 24 07:59:47 crc kubenswrapper[4809]: I1124 07:59:47.002175 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-mp9kw_781916ac-bfd7-4ce4-85c0-1e938fae3103/openstack-network-exporter/0.log" Nov 24 07:59:47 crc kubenswrapper[4809]: I1124 07:59:47.137840 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-drz48_3c2d3d10-789d-4a2d-9926-cba915a127df/ovsdb-server-init/0.log" Nov 24 07:59:47 crc kubenswrapper[4809]: I1124 07:59:47.150631 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_d9b4155b-3ac1-4c8d-b768-e28485799816/nova-metadata-metadata/0.log" Nov 24 07:59:47 crc kubenswrapper[4809]: I1124 07:59:47.373935 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-drz48_3c2d3d10-789d-4a2d-9926-cba915a127df/ovsdb-server-init/0.log" Nov 24 07:59:47 crc kubenswrapper[4809]: I1124 07:59:47.420522 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-drz48_3c2d3d10-789d-4a2d-9926-cba915a127df/ovs-vswitchd/0.log" Nov 24 07:59:47 crc kubenswrapper[4809]: I1124 07:59:47.442594 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-drz48_3c2d3d10-789d-4a2d-9926-cba915a127df/ovsdb-server/0.log" Nov 24 07:59:47 crc kubenswrapper[4809]: I1124 07:59:47.568833 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-vnpwc_c14ae567-41ab-455a-8756-04daf7c81677/ovn-controller/0.log" Nov 24 07:59:47 crc kubenswrapper[4809]: I1124 07:59:47.655981 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-mfjkm_a25b9532-a2c1-4e47-b8c2-7c5cec0fa255/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:59:47 crc kubenswrapper[4809]: I1124 07:59:47.810617 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_acc130b5-9141-4d3a-aec1-7778963967b0/openstack-network-exporter/0.log" Nov 24 07:59:47 crc kubenswrapper[4809]: I1124 07:59:47.901736 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_acc130b5-9141-4d3a-aec1-7778963967b0/ovn-northd/0.log" Nov 24 07:59:47 crc kubenswrapper[4809]: I1124 07:59:47.944692 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4be21c61-834b-4cff-bae7-ec458991ac7c/openstack-network-exporter/0.log" Nov 24 07:59:48 crc kubenswrapper[4809]: I1124 07:59:48.017826 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4be21c61-834b-4cff-bae7-ec458991ac7c/ovsdbserver-nb/0.log" Nov 24 07:59:48 crc kubenswrapper[4809]: I1124 07:59:48.137887 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe/openstack-network-exporter/0.log" Nov 24 07:59:48 crc kubenswrapper[4809]: I1124 07:59:48.263914 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ce74b3ec-a6ef-4fc4-a328-6a828b8f9afe/ovsdbserver-sb/0.log" Nov 24 07:59:48 crc kubenswrapper[4809]: I1124 07:59:48.415921 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-78fc658c7d-ntbgd_446a68c7-316d-43a1-af00-feba9b01e2dd/placement-api/0.log" Nov 24 07:59:48 crc kubenswrapper[4809]: I1124 07:59:48.500663 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-78fc658c7d-ntbgd_446a68c7-316d-43a1-af00-feba9b01e2dd/placement-log/0.log" Nov 24 07:59:48 crc kubenswrapper[4809]: I1124 07:59:48.505078 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3285fa33-509a-4959-8547-0332857328c9/setup-container/0.log" Nov 24 07:59:48 crc kubenswrapper[4809]: I1124 07:59:48.756419 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3285fa33-509a-4959-8547-0332857328c9/rabbitmq/0.log" Nov 24 07:59:48 crc kubenswrapper[4809]: I1124 07:59:48.761366 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3285fa33-509a-4959-8547-0332857328c9/setup-container/0.log" Nov 24 07:59:48 crc kubenswrapper[4809]: I1124 07:59:48.833576 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_4a1ec853-3d6a-403a-b2fe-82dc89ac5779/setup-container/0.log" Nov 24 07:59:49 crc kubenswrapper[4809]: I1124 07:59:49.014661 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_4a1ec853-3d6a-403a-b2fe-82dc89ac5779/setup-container/0.log" Nov 24 07:59:49 crc kubenswrapper[4809]: I1124 07:59:49.808634 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_4a1ec853-3d6a-403a-b2fe-82dc89ac5779/rabbitmq/0.log" Nov 24 07:59:50 crc kubenswrapper[4809]: I1124 07:59:50.097982 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-74rb2_a46047d8-9af2-4172-8e55-26dd29b587aa/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:59:50 crc kubenswrapper[4809]: I1124 07:59:50.251925 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-wmc74_7883d395-4314-4e1c-9a89-89b949d1b7a6/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:59:50 crc kubenswrapper[4809]: I1124 07:59:50.327016 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-m5nm9_9ed9735b-66b5-44b3-9399-af938bfcd802/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:59:50 crc kubenswrapper[4809]: I1124 07:59:50.484757 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-2fc45_2a469d7b-c3d8-46cf-b6cf-7fe13619f922/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:59:50 crc kubenswrapper[4809]: I1124 07:59:50.599570 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-ckss4_801a6c69-8f03-4f88-ba22-cd68875e8ae2/ssh-known-hosts-edpm-deployment/0.log" Nov 24 07:59:50 crc kubenswrapper[4809]: I1124 07:59:50.820509 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6d74d65d8c-rgmvx_2c1fe2a7-0b4f-46ee-8368-624499f8c095/proxy-server/0.log" Nov 24 07:59:50 crc kubenswrapper[4809]: I1124 07:59:50.864899 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-8n9sj_9ad3cf13-9872-497f-b8ee-180ec8d7bbdd/swift-ring-rebalance/0.log" Nov 24 07:59:50 crc kubenswrapper[4809]: I1124 07:59:50.897738 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6d74d65d8c-rgmvx_2c1fe2a7-0b4f-46ee-8368-624499f8c095/proxy-httpd/0.log" Nov 24 07:59:51 crc kubenswrapper[4809]: I1124 07:59:51.078750 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/account-auditor/0.log" Nov 24 07:59:51 crc kubenswrapper[4809]: I1124 07:59:51.116925 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/account-reaper/0.log" Nov 24 07:59:51 crc kubenswrapper[4809]: I1124 07:59:51.196743 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/account-replicator/0.log" Nov 24 07:59:51 crc kubenswrapper[4809]: I1124 07:59:51.299068 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/container-auditor/0.log" Nov 24 07:59:51 crc kubenswrapper[4809]: I1124 07:59:51.306339 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/account-server/0.log" Nov 24 07:59:51 crc kubenswrapper[4809]: I1124 07:59:51.415845 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/container-replicator/0.log" Nov 24 07:59:51 crc kubenswrapper[4809]: I1124 07:59:51.432140 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/container-server/0.log" Nov 24 07:59:51 crc kubenswrapper[4809]: I1124 07:59:51.525920 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/object-auditor/0.log" Nov 24 07:59:51 crc kubenswrapper[4809]: I1124 07:59:51.532160 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/container-updater/0.log" Nov 24 07:59:51 crc kubenswrapper[4809]: I1124 07:59:51.663478 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/object-replicator/0.log" Nov 24 07:59:51 crc kubenswrapper[4809]: I1124 07:59:51.688332 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/object-expirer/0.log" Nov 24 07:59:51 crc kubenswrapper[4809]: I1124 07:59:51.739530 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/object-server/0.log" Nov 24 07:59:51 crc kubenswrapper[4809]: I1124 07:59:51.761932 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/object-updater/0.log" Nov 24 07:59:51 crc kubenswrapper[4809]: I1124 07:59:51.868373 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/rsync/0.log" Nov 24 07:59:51 crc kubenswrapper[4809]: I1124 07:59:51.933861 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_53b09c3d-ced2-431c-ab5f-2c2e09c1a686/swift-recon-cron/0.log" Nov 24 07:59:52 crc kubenswrapper[4809]: I1124 07:59:52.037419 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-xgrwd_797b01f0-7757-495a-ad23-3973353832c1/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:59:52 crc kubenswrapper[4809]: I1124 07:59:52.206273 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_4f90573b-06dc-47d5-ab8b-5dab5503e96a/tempest-tests-tempest-tests-runner/0.log" Nov 24 07:59:52 crc kubenswrapper[4809]: I1124 07:59:52.208833 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_19e09f12-9159-4150-ae5b-6d718e6f9957/test-operator-logs-container/0.log" Nov 24 07:59:52 crc kubenswrapper[4809]: I1124 07:59:52.578177 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-mgk4f_2b6216ef-0d9c-4279-bf1c-526e2fe2d0dd/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 07:59:55 crc kubenswrapper[4809]: I1124 07:59:55.891948 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 07:59:56 crc kubenswrapper[4809]: I1124 07:59:56.528755 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerStarted","Data":"28e197e5ea01aea1c7445a8b46230eb9b3683e3bff6bd97e100500700dabc0ed"} Nov 24 08:00:00 crc kubenswrapper[4809]: I1124 08:00:00.154753 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399520-9fs2w"] Nov 24 08:00:00 crc kubenswrapper[4809]: E1124 08:00:00.155714 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c6307ff-83c1-44b4-badd-367d3dad9d51" containerName="container-00" Nov 24 08:00:00 crc kubenswrapper[4809]: I1124 08:00:00.155725 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c6307ff-83c1-44b4-badd-367d3dad9d51" containerName="container-00" Nov 24 08:00:00 crc kubenswrapper[4809]: I1124 08:00:00.155906 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c6307ff-83c1-44b4-badd-367d3dad9d51" containerName="container-00" Nov 24 08:00:00 crc kubenswrapper[4809]: I1124 08:00:00.156551 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-9fs2w" Nov 24 08:00:00 crc kubenswrapper[4809]: I1124 08:00:00.166780 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399520-9fs2w"] Nov 24 08:00:00 crc kubenswrapper[4809]: I1124 08:00:00.180844 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 08:00:00 crc kubenswrapper[4809]: I1124 08:00:00.180939 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 08:00:00 crc kubenswrapper[4809]: I1124 08:00:00.189888 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xxm6\" (UniqueName: \"kubernetes.io/projected/13606173-b82a-4b88-b5ed-bba5c243d486-kube-api-access-5xxm6\") pod \"collect-profiles-29399520-9fs2w\" (UID: \"13606173-b82a-4b88-b5ed-bba5c243d486\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-9fs2w" Nov 24 08:00:00 crc kubenswrapper[4809]: I1124 08:00:00.190041 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/13606173-b82a-4b88-b5ed-bba5c243d486-secret-volume\") pod \"collect-profiles-29399520-9fs2w\" (UID: \"13606173-b82a-4b88-b5ed-bba5c243d486\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-9fs2w" Nov 24 08:00:00 crc kubenswrapper[4809]: I1124 08:00:00.190087 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/13606173-b82a-4b88-b5ed-bba5c243d486-config-volume\") pod \"collect-profiles-29399520-9fs2w\" (UID: \"13606173-b82a-4b88-b5ed-bba5c243d486\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-9fs2w" Nov 24 08:00:00 crc kubenswrapper[4809]: I1124 08:00:00.291491 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/13606173-b82a-4b88-b5ed-bba5c243d486-secret-volume\") pod \"collect-profiles-29399520-9fs2w\" (UID: \"13606173-b82a-4b88-b5ed-bba5c243d486\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-9fs2w" Nov 24 08:00:00 crc kubenswrapper[4809]: I1124 08:00:00.291559 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/13606173-b82a-4b88-b5ed-bba5c243d486-config-volume\") pod \"collect-profiles-29399520-9fs2w\" (UID: \"13606173-b82a-4b88-b5ed-bba5c243d486\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-9fs2w" Nov 24 08:00:00 crc kubenswrapper[4809]: I1124 08:00:00.291619 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xxm6\" (UniqueName: \"kubernetes.io/projected/13606173-b82a-4b88-b5ed-bba5c243d486-kube-api-access-5xxm6\") pod \"collect-profiles-29399520-9fs2w\" (UID: \"13606173-b82a-4b88-b5ed-bba5c243d486\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-9fs2w" Nov 24 08:00:00 crc kubenswrapper[4809]: I1124 08:00:00.292797 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/13606173-b82a-4b88-b5ed-bba5c243d486-config-volume\") pod \"collect-profiles-29399520-9fs2w\" (UID: \"13606173-b82a-4b88-b5ed-bba5c243d486\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-9fs2w" Nov 24 08:00:00 crc kubenswrapper[4809]: I1124 08:00:00.303056 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/13606173-b82a-4b88-b5ed-bba5c243d486-secret-volume\") pod \"collect-profiles-29399520-9fs2w\" (UID: \"13606173-b82a-4b88-b5ed-bba5c243d486\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-9fs2w" Nov 24 08:00:00 crc kubenswrapper[4809]: I1124 08:00:00.320743 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xxm6\" (UniqueName: \"kubernetes.io/projected/13606173-b82a-4b88-b5ed-bba5c243d486-kube-api-access-5xxm6\") pod \"collect-profiles-29399520-9fs2w\" (UID: \"13606173-b82a-4b88-b5ed-bba5c243d486\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-9fs2w" Nov 24 08:00:00 crc kubenswrapper[4809]: I1124 08:00:00.505076 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-9fs2w" Nov 24 08:00:00 crc kubenswrapper[4809]: I1124 08:00:00.965273 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399520-9fs2w"] Nov 24 08:00:01 crc kubenswrapper[4809]: I1124 08:00:01.586636 4809 generic.go:334] "Generic (PLEG): container finished" podID="13606173-b82a-4b88-b5ed-bba5c243d486" containerID="c3c1ccf57524d875c0b5fcd1bc5b871632ec121b0c993e517f1b08668d1c33d7" exitCode=0 Nov 24 08:00:01 crc kubenswrapper[4809]: I1124 08:00:01.587054 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-9fs2w" event={"ID":"13606173-b82a-4b88-b5ed-bba5c243d486","Type":"ContainerDied","Data":"c3c1ccf57524d875c0b5fcd1bc5b871632ec121b0c993e517f1b08668d1c33d7"} Nov 24 08:00:01 crc kubenswrapper[4809]: I1124 08:00:01.587082 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-9fs2w" event={"ID":"13606173-b82a-4b88-b5ed-bba5c243d486","Type":"ContainerStarted","Data":"6535894b1979909504582d9d223f3a6052f3c740813e6e0c0316ede14bfdad2f"} Nov 24 08:00:01 crc kubenswrapper[4809]: I1124 08:00:01.590231 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_ced578a4-f9b1-4d1f-978f-beda4b33bc59/memcached/0.log" Nov 24 08:00:02 crc kubenswrapper[4809]: I1124 08:00:02.957411 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-9fs2w" Nov 24 08:00:03 crc kubenswrapper[4809]: I1124 08:00:03.059277 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/13606173-b82a-4b88-b5ed-bba5c243d486-config-volume\") pod \"13606173-b82a-4b88-b5ed-bba5c243d486\" (UID: \"13606173-b82a-4b88-b5ed-bba5c243d486\") " Nov 24 08:00:03 crc kubenswrapper[4809]: I1124 08:00:03.060722 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/13606173-b82a-4b88-b5ed-bba5c243d486-secret-volume\") pod \"13606173-b82a-4b88-b5ed-bba5c243d486\" (UID: \"13606173-b82a-4b88-b5ed-bba5c243d486\") " Nov 24 08:00:03 crc kubenswrapper[4809]: I1124 08:00:03.060800 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xxm6\" (UniqueName: \"kubernetes.io/projected/13606173-b82a-4b88-b5ed-bba5c243d486-kube-api-access-5xxm6\") pod \"13606173-b82a-4b88-b5ed-bba5c243d486\" (UID: \"13606173-b82a-4b88-b5ed-bba5c243d486\") " Nov 24 08:00:03 crc kubenswrapper[4809]: I1124 08:00:03.060566 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13606173-b82a-4b88-b5ed-bba5c243d486-config-volume" (OuterVolumeSpecName: "config-volume") pod "13606173-b82a-4b88-b5ed-bba5c243d486" (UID: "13606173-b82a-4b88-b5ed-bba5c243d486"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:00:03 crc kubenswrapper[4809]: I1124 08:00:03.061555 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/13606173-b82a-4b88-b5ed-bba5c243d486-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 08:00:03 crc kubenswrapper[4809]: I1124 08:00:03.071574 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13606173-b82a-4b88-b5ed-bba5c243d486-kube-api-access-5xxm6" (OuterVolumeSpecName: "kube-api-access-5xxm6") pod "13606173-b82a-4b88-b5ed-bba5c243d486" (UID: "13606173-b82a-4b88-b5ed-bba5c243d486"). InnerVolumeSpecName "kube-api-access-5xxm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:00:03 crc kubenswrapper[4809]: I1124 08:00:03.071870 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13606173-b82a-4b88-b5ed-bba5c243d486-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "13606173-b82a-4b88-b5ed-bba5c243d486" (UID: "13606173-b82a-4b88-b5ed-bba5c243d486"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:00:03 crc kubenswrapper[4809]: I1124 08:00:03.163681 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/13606173-b82a-4b88-b5ed-bba5c243d486-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 08:00:03 crc kubenswrapper[4809]: I1124 08:00:03.163723 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xxm6\" (UniqueName: \"kubernetes.io/projected/13606173-b82a-4b88-b5ed-bba5c243d486-kube-api-access-5xxm6\") on node \"crc\" DevicePath \"\"" Nov 24 08:00:03 crc kubenswrapper[4809]: I1124 08:00:03.606536 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-9fs2w" event={"ID":"13606173-b82a-4b88-b5ed-bba5c243d486","Type":"ContainerDied","Data":"6535894b1979909504582d9d223f3a6052f3c740813e6e0c0316ede14bfdad2f"} Nov 24 08:00:03 crc kubenswrapper[4809]: I1124 08:00:03.606575 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6535894b1979909504582d9d223f3a6052f3c740813e6e0c0316ede14bfdad2f" Nov 24 08:00:03 crc kubenswrapper[4809]: I1124 08:00:03.606587 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-9fs2w" Nov 24 08:00:04 crc kubenswrapper[4809]: I1124 08:00:04.042843 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399475-h4sk5"] Nov 24 08:00:04 crc kubenswrapper[4809]: I1124 08:00:04.051857 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399475-h4sk5"] Nov 24 08:00:04 crc kubenswrapper[4809]: I1124 08:00:04.917661 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b19ae12c-c3c5-4082-be4f-ffde1d1f56f0" path="/var/lib/kubelet/pods/b19ae12c-c3c5-4082-be4f-ffde1d1f56f0/volumes" Nov 24 08:00:11 crc kubenswrapper[4809]: I1124 08:00:11.041085 4809 scope.go:117] "RemoveContainer" containerID="093604f5c39be5a54b49888cf4e0f869d9d6902fc20cb31eb8fccf902e3574d2" Nov 24 08:00:16 crc kubenswrapper[4809]: I1124 08:00:16.126095 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b_ed47ae85-6bd2-4e58-acee-30c9022b9170/util/0.log" Nov 24 08:00:16 crc kubenswrapper[4809]: I1124 08:00:16.277451 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b_ed47ae85-6bd2-4e58-acee-30c9022b9170/pull/0.log" Nov 24 08:00:16 crc kubenswrapper[4809]: I1124 08:00:16.305701 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b_ed47ae85-6bd2-4e58-acee-30c9022b9170/util/0.log" Nov 24 08:00:16 crc kubenswrapper[4809]: I1124 08:00:16.347808 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b_ed47ae85-6bd2-4e58-acee-30c9022b9170/pull/0.log" Nov 24 08:00:16 crc kubenswrapper[4809]: I1124 08:00:16.526019 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b_ed47ae85-6bd2-4e58-acee-30c9022b9170/extract/0.log" Nov 24 08:00:16 crc kubenswrapper[4809]: I1124 08:00:16.528309 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b_ed47ae85-6bd2-4e58-acee-30c9022b9170/util/0.log" Nov 24 08:00:16 crc kubenswrapper[4809]: I1124 08:00:16.543588 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_33033ec3afac6662211b402fbe53dfb4a0d0ed3237d37e2ebae5817c505cb2b_ed47ae85-6bd2-4e58-acee-30c9022b9170/pull/0.log" Nov 24 08:00:16 crc kubenswrapper[4809]: I1124 08:00:16.711524 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-6k9xr_bfcf902b-83f0-48f9-ad6b-eb7963f43af4/kube-rbac-proxy/0.log" Nov 24 08:00:16 crc kubenswrapper[4809]: I1124 08:00:16.723005 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-49dpw_e525562e-b6b2-415f-b6e7-1c0bfde8a56f/kube-rbac-proxy/0.log" Nov 24 08:00:16 crc kubenswrapper[4809]: I1124 08:00:16.757540 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-6k9xr_bfcf902b-83f0-48f9-ad6b-eb7963f43af4/manager/0.log" Nov 24 08:00:16 crc kubenswrapper[4809]: I1124 08:00:16.904726 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-49dpw_e525562e-b6b2-415f-b6e7-1c0bfde8a56f/manager/0.log" Nov 24 08:00:16 crc kubenswrapper[4809]: I1124 08:00:16.948884 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-plhlb_200f6896-3100-46ea-94f4-69637762a259/kube-rbac-proxy/0.log" Nov 24 08:00:16 crc kubenswrapper[4809]: I1124 08:00:16.966201 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-plhlb_200f6896-3100-46ea-94f4-69637762a259/manager/0.log" Nov 24 08:00:17 crc kubenswrapper[4809]: I1124 08:00:17.141587 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-8nzmk_0a1424d6-9ecf-4a06-b174-15d13f61b916/kube-rbac-proxy/0.log" Nov 24 08:00:17 crc kubenswrapper[4809]: I1124 08:00:17.164529 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-8nzmk_0a1424d6-9ecf-4a06-b174-15d13f61b916/manager/0.log" Nov 24 08:00:17 crc kubenswrapper[4809]: I1124 08:00:17.279092 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-4ddch_d78cc4fe-bab9-4521-a1d2-ed6efd341789/kube-rbac-proxy/0.log" Nov 24 08:00:17 crc kubenswrapper[4809]: I1124 08:00:17.313060 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-4ddch_d78cc4fe-bab9-4521-a1d2-ed6efd341789/manager/0.log" Nov 24 08:00:17 crc kubenswrapper[4809]: I1124 08:00:17.372408 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-xf4fh_4d4aa35c-faa7-4b93-aecd-ee4645f11916/kube-rbac-proxy/0.log" Nov 24 08:00:17 crc kubenswrapper[4809]: I1124 08:00:17.485480 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-xf4fh_4d4aa35c-faa7-4b93-aecd-ee4645f11916/manager/0.log" Nov 24 08:00:17 crc kubenswrapper[4809]: I1124 08:00:17.594992 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-4vrfb_5b90cc28-7c58-4826-b849-3456bb654f0d/kube-rbac-proxy/0.log" Nov 24 08:00:17 crc kubenswrapper[4809]: I1124 08:00:17.713825 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-4vrfb_5b90cc28-7c58-4826-b849-3456bb654f0d/manager/0.log" Nov 24 08:00:17 crc kubenswrapper[4809]: I1124 08:00:17.729339 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-jfjqt_5f4f9ac0-6714-4f4a-a7fb-8d65b19c919c/kube-rbac-proxy/0.log" Nov 24 08:00:17 crc kubenswrapper[4809]: I1124 08:00:17.769659 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-jfjqt_5f4f9ac0-6714-4f4a-a7fb-8d65b19c919c/manager/0.log" Nov 24 08:00:17 crc kubenswrapper[4809]: I1124 08:00:17.884174 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-wh5fq_b0a19ac9-d266-4416-8590-3c8de4dadc7f/kube-rbac-proxy/0.log" Nov 24 08:00:18 crc kubenswrapper[4809]: I1124 08:00:18.000894 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-wh5fq_b0a19ac9-d266-4416-8590-3c8de4dadc7f/manager/0.log" Nov 24 08:00:18 crc kubenswrapper[4809]: I1124 08:00:18.092234 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-bz8c8_a2b969b3-07c7-4bf3-ab18-5317d8f0a372/kube-rbac-proxy/0.log" Nov 24 08:00:18 crc kubenswrapper[4809]: I1124 08:00:18.122313 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-bz8c8_a2b969b3-07c7-4bf3-ab18-5317d8f0a372/manager/0.log" Nov 24 08:00:18 crc kubenswrapper[4809]: I1124 08:00:18.202784 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-rp7r4_2f364be5-58b1-4d57-97f5-65a13d63e34e/kube-rbac-proxy/0.log" Nov 24 08:00:18 crc kubenswrapper[4809]: I1124 08:00:18.262514 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-rp7r4_2f364be5-58b1-4d57-97f5-65a13d63e34e/manager/0.log" Nov 24 08:00:18 crc kubenswrapper[4809]: I1124 08:00:18.359875 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-rgkd9_e94d989e-7aee-43d9-87c0-712355ca193f/kube-rbac-proxy/0.log" Nov 24 08:00:18 crc kubenswrapper[4809]: I1124 08:00:18.423521 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-rgkd9_e94d989e-7aee-43d9-87c0-712355ca193f/manager/0.log" Nov 24 08:00:18 crc kubenswrapper[4809]: I1124 08:00:18.534988 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-tk2hk_fd9ddd03-bc95-43e1-93ef-97ad5ca589b3/kube-rbac-proxy/0.log" Nov 24 08:00:18 crc kubenswrapper[4809]: I1124 08:00:18.607281 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-tk2hk_fd9ddd03-bc95-43e1-93ef-97ad5ca589b3/manager/0.log" Nov 24 08:00:18 crc kubenswrapper[4809]: I1124 08:00:18.635916 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-84nbm_bfe0f53e-685b-4cef-b763-90399c5fdaac/kube-rbac-proxy/0.log" Nov 24 08:00:18 crc kubenswrapper[4809]: I1124 08:00:18.753115 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-84nbm_bfe0f53e-685b-4cef-b763-90399c5fdaac/manager/0.log" Nov 24 08:00:18 crc kubenswrapper[4809]: I1124 08:00:18.787342 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-7vg75_2e599303-b44b-4ce5-9ee9-1e1c13ca856f/kube-rbac-proxy/0.log" Nov 24 08:00:18 crc kubenswrapper[4809]: I1124 08:00:18.846316 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-7vg75_2e599303-b44b-4ce5-9ee9-1e1c13ca856f/manager/0.log" Nov 24 08:00:18 crc kubenswrapper[4809]: I1124 08:00:18.981941 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8fcb84d5c-xqd67_5b38e6fb-1aef-4ecd-80ca-4665a9d3430a/kube-rbac-proxy/0.log" Nov 24 08:00:19 crc kubenswrapper[4809]: I1124 08:00:19.239696 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6c94b45f4d-pdjb4_a9cc582d-d6ce-4bb6-b698-9733ce8be925/kube-rbac-proxy/0.log" Nov 24 08:00:19 crc kubenswrapper[4809]: I1124 08:00:19.390889 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6c94b45f4d-pdjb4_a9cc582d-d6ce-4bb6-b698-9733ce8be925/operator/0.log" Nov 24 08:00:19 crc kubenswrapper[4809]: I1124 08:00:19.429111 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-mhj4l_bb31150b-b8eb-4340-ba08-384de196289a/registry-server/0.log" Nov 24 08:00:19 crc kubenswrapper[4809]: I1124 08:00:19.611712 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-59brn_9eeb9814-10f8-478d-888f-5cf058c99c4b/kube-rbac-proxy/0.log" Nov 24 08:00:19 crc kubenswrapper[4809]: I1124 08:00:19.710181 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-59brn_9eeb9814-10f8-478d-888f-5cf058c99c4b/manager/0.log" Nov 24 08:00:19 crc kubenswrapper[4809]: I1124 08:00:19.812071 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-rtxsf_483b5843-7751-465a-9412-dbb15ef5145f/kube-rbac-proxy/0.log" Nov 24 08:00:20 crc kubenswrapper[4809]: I1124 08:00:20.013003 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8fcb84d5c-xqd67_5b38e6fb-1aef-4ecd-80ca-4665a9d3430a/manager/0.log" Nov 24 08:00:20 crc kubenswrapper[4809]: I1124 08:00:20.133696 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-px9vl_02a1aa7c-c3f7-4dcb-89ad-8d34f46ff4a1/operator/0.log" Nov 24 08:00:20 crc kubenswrapper[4809]: I1124 08:00:20.158120 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-rtxsf_483b5843-7751-465a-9412-dbb15ef5145f/manager/0.log" Nov 24 08:00:20 crc kubenswrapper[4809]: I1124 08:00:20.258716 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-mfgld_6269a5dc-fd2a-4fdd-9b18-4e505242efca/kube-rbac-proxy/0.log" Nov 24 08:00:20 crc kubenswrapper[4809]: I1124 08:00:20.303740 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-mfgld_6269a5dc-fd2a-4fdd-9b18-4e505242efca/manager/0.log" Nov 24 08:00:20 crc kubenswrapper[4809]: I1124 08:00:20.363781 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6d4bf84b58-4xjsj_39fc255b-7ac9-40b4-98e0-74d1f76c2696/kube-rbac-proxy/0.log" Nov 24 08:00:20 crc kubenswrapper[4809]: I1124 08:00:20.474242 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6d4bf84b58-4xjsj_39fc255b-7ac9-40b4-98e0-74d1f76c2696/manager/0.log" Nov 24 08:00:20 crc kubenswrapper[4809]: I1124 08:00:20.478046 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-wbzcn_cb240075-d846-4751-b026-6fa452a565a5/kube-rbac-proxy/0.log" Nov 24 08:00:20 crc kubenswrapper[4809]: I1124 08:00:20.531375 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-wbzcn_cb240075-d846-4751-b026-6fa452a565a5/manager/0.log" Nov 24 08:00:20 crc kubenswrapper[4809]: I1124 08:00:20.630224 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-twjz4_df289b0d-b465-49d2-b1ee-4f0faf87b079/manager/0.log" Nov 24 08:00:20 crc kubenswrapper[4809]: I1124 08:00:20.659586 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-twjz4_df289b0d-b465-49d2-b1ee-4f0faf87b079/kube-rbac-proxy/0.log" Nov 24 08:00:27 crc kubenswrapper[4809]: I1124 08:00:27.091055 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l9bcn"] Nov 24 08:00:27 crc kubenswrapper[4809]: E1124 08:00:27.092143 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13606173-b82a-4b88-b5ed-bba5c243d486" containerName="collect-profiles" Nov 24 08:00:27 crc kubenswrapper[4809]: I1124 08:00:27.092161 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="13606173-b82a-4b88-b5ed-bba5c243d486" containerName="collect-profiles" Nov 24 08:00:27 crc kubenswrapper[4809]: I1124 08:00:27.092403 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="13606173-b82a-4b88-b5ed-bba5c243d486" containerName="collect-profiles" Nov 24 08:00:27 crc kubenswrapper[4809]: I1124 08:00:27.094162 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9bcn" Nov 24 08:00:27 crc kubenswrapper[4809]: I1124 08:00:27.099730 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l9bcn"] Nov 24 08:00:27 crc kubenswrapper[4809]: I1124 08:00:27.199857 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z92rx\" (UniqueName: \"kubernetes.io/projected/0f303785-0349-4fe0-b300-19efa68d61c4-kube-api-access-z92rx\") pod \"redhat-operators-l9bcn\" (UID: \"0f303785-0349-4fe0-b300-19efa68d61c4\") " pod="openshift-marketplace/redhat-operators-l9bcn" Nov 24 08:00:27 crc kubenswrapper[4809]: I1124 08:00:27.199917 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f303785-0349-4fe0-b300-19efa68d61c4-utilities\") pod \"redhat-operators-l9bcn\" (UID: \"0f303785-0349-4fe0-b300-19efa68d61c4\") " pod="openshift-marketplace/redhat-operators-l9bcn" Nov 24 08:00:27 crc kubenswrapper[4809]: I1124 08:00:27.200047 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f303785-0349-4fe0-b300-19efa68d61c4-catalog-content\") pod \"redhat-operators-l9bcn\" (UID: \"0f303785-0349-4fe0-b300-19efa68d61c4\") " pod="openshift-marketplace/redhat-operators-l9bcn" Nov 24 08:00:27 crc kubenswrapper[4809]: I1124 08:00:27.302378 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z92rx\" (UniqueName: \"kubernetes.io/projected/0f303785-0349-4fe0-b300-19efa68d61c4-kube-api-access-z92rx\") pod \"redhat-operators-l9bcn\" (UID: \"0f303785-0349-4fe0-b300-19efa68d61c4\") " pod="openshift-marketplace/redhat-operators-l9bcn" Nov 24 08:00:27 crc kubenswrapper[4809]: I1124 08:00:27.302441 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f303785-0349-4fe0-b300-19efa68d61c4-utilities\") pod \"redhat-operators-l9bcn\" (UID: \"0f303785-0349-4fe0-b300-19efa68d61c4\") " pod="openshift-marketplace/redhat-operators-l9bcn" Nov 24 08:00:27 crc kubenswrapper[4809]: I1124 08:00:27.302550 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f303785-0349-4fe0-b300-19efa68d61c4-catalog-content\") pod \"redhat-operators-l9bcn\" (UID: \"0f303785-0349-4fe0-b300-19efa68d61c4\") " pod="openshift-marketplace/redhat-operators-l9bcn" Nov 24 08:00:27 crc kubenswrapper[4809]: I1124 08:00:27.303204 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f303785-0349-4fe0-b300-19efa68d61c4-catalog-content\") pod \"redhat-operators-l9bcn\" (UID: \"0f303785-0349-4fe0-b300-19efa68d61c4\") " pod="openshift-marketplace/redhat-operators-l9bcn" Nov 24 08:00:27 crc kubenswrapper[4809]: I1124 08:00:27.303200 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f303785-0349-4fe0-b300-19efa68d61c4-utilities\") pod \"redhat-operators-l9bcn\" (UID: \"0f303785-0349-4fe0-b300-19efa68d61c4\") " pod="openshift-marketplace/redhat-operators-l9bcn" Nov 24 08:00:27 crc kubenswrapper[4809]: I1124 08:00:27.341764 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z92rx\" (UniqueName: \"kubernetes.io/projected/0f303785-0349-4fe0-b300-19efa68d61c4-kube-api-access-z92rx\") pod \"redhat-operators-l9bcn\" (UID: \"0f303785-0349-4fe0-b300-19efa68d61c4\") " pod="openshift-marketplace/redhat-operators-l9bcn" Nov 24 08:00:27 crc kubenswrapper[4809]: I1124 08:00:27.420210 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9bcn" Nov 24 08:00:27 crc kubenswrapper[4809]: I1124 08:00:27.892213 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l9bcn"] Nov 24 08:00:28 crc kubenswrapper[4809]: I1124 08:00:28.824688 4809 generic.go:334] "Generic (PLEG): container finished" podID="0f303785-0349-4fe0-b300-19efa68d61c4" containerID="43b3c9ed4165a96f89aab8fdd98c98aaae3d8391dd5fe97e6ded3c4dbe3a6630" exitCode=0 Nov 24 08:00:28 crc kubenswrapper[4809]: I1124 08:00:28.824772 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9bcn" event={"ID":"0f303785-0349-4fe0-b300-19efa68d61c4","Type":"ContainerDied","Data":"43b3c9ed4165a96f89aab8fdd98c98aaae3d8391dd5fe97e6ded3c4dbe3a6630"} Nov 24 08:00:28 crc kubenswrapper[4809]: I1124 08:00:28.825062 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9bcn" event={"ID":"0f303785-0349-4fe0-b300-19efa68d61c4","Type":"ContainerStarted","Data":"ec45d8bed1a3d5973a2d6fa32a6a835586c56dc1d6225516587ab1dfad3dcd56"} Nov 24 08:00:28 crc kubenswrapper[4809]: I1124 08:00:28.828444 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 08:00:29 crc kubenswrapper[4809]: I1124 08:00:29.835444 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9bcn" event={"ID":"0f303785-0349-4fe0-b300-19efa68d61c4","Type":"ContainerStarted","Data":"1ade393e694b53bd26850e9bed1d6dc98d9a2a57d846fbbe53cdaa1972b3e32c"} Nov 24 08:00:33 crc kubenswrapper[4809]: I1124 08:00:33.871798 4809 generic.go:334] "Generic (PLEG): container finished" podID="0f303785-0349-4fe0-b300-19efa68d61c4" containerID="1ade393e694b53bd26850e9bed1d6dc98d9a2a57d846fbbe53cdaa1972b3e32c" exitCode=0 Nov 24 08:00:33 crc kubenswrapper[4809]: I1124 08:00:33.871867 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9bcn" event={"ID":"0f303785-0349-4fe0-b300-19efa68d61c4","Type":"ContainerDied","Data":"1ade393e694b53bd26850e9bed1d6dc98d9a2a57d846fbbe53cdaa1972b3e32c"} Nov 24 08:00:34 crc kubenswrapper[4809]: I1124 08:00:34.882569 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9bcn" event={"ID":"0f303785-0349-4fe0-b300-19efa68d61c4","Type":"ContainerStarted","Data":"fde4c645dfb146fa707a4545d70e5bcf05ff688dee2e28b1711d1cfcedab2ef7"} Nov 24 08:00:34 crc kubenswrapper[4809]: I1124 08:00:34.909834 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l9bcn" podStartSLOduration=2.452123471 podStartE2EDuration="7.90980763s" podCreationTimestamp="2025-11-24 08:00:27 +0000 UTC" firstStartedPulling="2025-11-24 08:00:28.828182453 +0000 UTC m=+3928.728774158" lastFinishedPulling="2025-11-24 08:00:34.285866562 +0000 UTC m=+3934.186458317" observedRunningTime="2025-11-24 08:00:34.905052441 +0000 UTC m=+3934.805644146" watchObservedRunningTime="2025-11-24 08:00:34.90980763 +0000 UTC m=+3934.810399335" Nov 24 08:00:35 crc kubenswrapper[4809]: I1124 08:00:35.832711 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-xfdvj_3e14d56e-43b3-40ec-809f-371e8ab2f628/control-plane-machine-set-operator/0.log" Nov 24 08:00:36 crc kubenswrapper[4809]: I1124 08:00:36.044615 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-56j57_eca6947b-dd5e-449d-a8c5-05277c556bd1/kube-rbac-proxy/0.log" Nov 24 08:00:36 crc kubenswrapper[4809]: I1124 08:00:36.068205 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-56j57_eca6947b-dd5e-449d-a8c5-05277c556bd1/machine-api-operator/0.log" Nov 24 08:00:37 crc kubenswrapper[4809]: I1124 08:00:37.421330 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l9bcn" Nov 24 08:00:37 crc kubenswrapper[4809]: I1124 08:00:37.421380 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l9bcn" Nov 24 08:00:38 crc kubenswrapper[4809]: I1124 08:00:38.478298 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l9bcn" podUID="0f303785-0349-4fe0-b300-19efa68d61c4" containerName="registry-server" probeResult="failure" output=< Nov 24 08:00:38 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Nov 24 08:00:38 crc kubenswrapper[4809]: > Nov 24 08:00:47 crc kubenswrapper[4809]: I1124 08:00:47.481878 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l9bcn" Nov 24 08:00:47 crc kubenswrapper[4809]: I1124 08:00:47.549516 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l9bcn" Nov 24 08:00:47 crc kubenswrapper[4809]: I1124 08:00:47.716372 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l9bcn"] Nov 24 08:00:48 crc kubenswrapper[4809]: I1124 08:00:48.786151 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-hm272_1b39cb48-19bf-4ff6-a9e7-a2270b28ab87/cert-manager-controller/0.log" Nov 24 08:00:49 crc kubenswrapper[4809]: I1124 08:00:49.020281 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l9bcn" podUID="0f303785-0349-4fe0-b300-19efa68d61c4" containerName="registry-server" containerID="cri-o://fde4c645dfb146fa707a4545d70e5bcf05ff688dee2e28b1711d1cfcedab2ef7" gracePeriod=2 Nov 24 08:00:49 crc kubenswrapper[4809]: I1124 08:00:49.110521 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-wslw7_4e03214c-a15c-4a2f-a802-ce80e9fde1c8/cert-manager-webhook/0.log" Nov 24 08:00:49 crc kubenswrapper[4809]: I1124 08:00:49.206796 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-xhsz2_9de96055-2c3b-4012-a260-7b55cbb0df1b/cert-manager-cainjector/0.log" Nov 24 08:00:49 crc kubenswrapper[4809]: I1124 08:00:49.494858 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9bcn" Nov 24 08:00:49 crc kubenswrapper[4809]: I1124 08:00:49.618680 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z92rx\" (UniqueName: \"kubernetes.io/projected/0f303785-0349-4fe0-b300-19efa68d61c4-kube-api-access-z92rx\") pod \"0f303785-0349-4fe0-b300-19efa68d61c4\" (UID: \"0f303785-0349-4fe0-b300-19efa68d61c4\") " Nov 24 08:00:49 crc kubenswrapper[4809]: I1124 08:00:49.618765 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f303785-0349-4fe0-b300-19efa68d61c4-utilities\") pod \"0f303785-0349-4fe0-b300-19efa68d61c4\" (UID: \"0f303785-0349-4fe0-b300-19efa68d61c4\") " Nov 24 08:00:49 crc kubenswrapper[4809]: I1124 08:00:49.618929 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f303785-0349-4fe0-b300-19efa68d61c4-catalog-content\") pod \"0f303785-0349-4fe0-b300-19efa68d61c4\" (UID: \"0f303785-0349-4fe0-b300-19efa68d61c4\") " Nov 24 08:00:49 crc kubenswrapper[4809]: I1124 08:00:49.620388 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f303785-0349-4fe0-b300-19efa68d61c4-utilities" (OuterVolumeSpecName: "utilities") pod "0f303785-0349-4fe0-b300-19efa68d61c4" (UID: "0f303785-0349-4fe0-b300-19efa68d61c4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:00:49 crc kubenswrapper[4809]: I1124 08:00:49.625588 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f303785-0349-4fe0-b300-19efa68d61c4-kube-api-access-z92rx" (OuterVolumeSpecName: "kube-api-access-z92rx") pod "0f303785-0349-4fe0-b300-19efa68d61c4" (UID: "0f303785-0349-4fe0-b300-19efa68d61c4"). InnerVolumeSpecName "kube-api-access-z92rx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:00:49 crc kubenswrapper[4809]: I1124 08:00:49.720045 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f303785-0349-4fe0-b300-19efa68d61c4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0f303785-0349-4fe0-b300-19efa68d61c4" (UID: "0f303785-0349-4fe0-b300-19efa68d61c4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:00:49 crc kubenswrapper[4809]: I1124 08:00:49.721330 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z92rx\" (UniqueName: \"kubernetes.io/projected/0f303785-0349-4fe0-b300-19efa68d61c4-kube-api-access-z92rx\") on node \"crc\" DevicePath \"\"" Nov 24 08:00:49 crc kubenswrapper[4809]: I1124 08:00:49.721357 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f303785-0349-4fe0-b300-19efa68d61c4-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:00:49 crc kubenswrapper[4809]: I1124 08:00:49.721367 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f303785-0349-4fe0-b300-19efa68d61c4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:00:50 crc kubenswrapper[4809]: I1124 08:00:50.030496 4809 generic.go:334] "Generic (PLEG): container finished" podID="0f303785-0349-4fe0-b300-19efa68d61c4" containerID="fde4c645dfb146fa707a4545d70e5bcf05ff688dee2e28b1711d1cfcedab2ef7" exitCode=0 Nov 24 08:00:50 crc kubenswrapper[4809]: I1124 08:00:50.030541 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9bcn" event={"ID":"0f303785-0349-4fe0-b300-19efa68d61c4","Type":"ContainerDied","Data":"fde4c645dfb146fa707a4545d70e5bcf05ff688dee2e28b1711d1cfcedab2ef7"} Nov 24 08:00:50 crc kubenswrapper[4809]: I1124 08:00:50.030603 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9bcn" event={"ID":"0f303785-0349-4fe0-b300-19efa68d61c4","Type":"ContainerDied","Data":"ec45d8bed1a3d5973a2d6fa32a6a835586c56dc1d6225516587ab1dfad3dcd56"} Nov 24 08:00:50 crc kubenswrapper[4809]: I1124 08:00:50.030631 4809 scope.go:117] "RemoveContainer" containerID="fde4c645dfb146fa707a4545d70e5bcf05ff688dee2e28b1711d1cfcedab2ef7" Nov 24 08:00:50 crc kubenswrapper[4809]: I1124 08:00:50.031987 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9bcn" Nov 24 08:00:50 crc kubenswrapper[4809]: I1124 08:00:50.049372 4809 scope.go:117] "RemoveContainer" containerID="1ade393e694b53bd26850e9bed1d6dc98d9a2a57d846fbbe53cdaa1972b3e32c" Nov 24 08:00:50 crc kubenswrapper[4809]: I1124 08:00:50.068979 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l9bcn"] Nov 24 08:00:50 crc kubenswrapper[4809]: I1124 08:00:50.078746 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l9bcn"] Nov 24 08:00:50 crc kubenswrapper[4809]: I1124 08:00:50.092053 4809 scope.go:117] "RemoveContainer" containerID="43b3c9ed4165a96f89aab8fdd98c98aaae3d8391dd5fe97e6ded3c4dbe3a6630" Nov 24 08:00:50 crc kubenswrapper[4809]: I1124 08:00:50.123226 4809 scope.go:117] "RemoveContainer" containerID="fde4c645dfb146fa707a4545d70e5bcf05ff688dee2e28b1711d1cfcedab2ef7" Nov 24 08:00:50 crc kubenswrapper[4809]: E1124 08:00:50.127605 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fde4c645dfb146fa707a4545d70e5bcf05ff688dee2e28b1711d1cfcedab2ef7\": container with ID starting with fde4c645dfb146fa707a4545d70e5bcf05ff688dee2e28b1711d1cfcedab2ef7 not found: ID does not exist" containerID="fde4c645dfb146fa707a4545d70e5bcf05ff688dee2e28b1711d1cfcedab2ef7" Nov 24 08:00:50 crc kubenswrapper[4809]: I1124 08:00:50.127654 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fde4c645dfb146fa707a4545d70e5bcf05ff688dee2e28b1711d1cfcedab2ef7"} err="failed to get container status \"fde4c645dfb146fa707a4545d70e5bcf05ff688dee2e28b1711d1cfcedab2ef7\": rpc error: code = NotFound desc = could not find container \"fde4c645dfb146fa707a4545d70e5bcf05ff688dee2e28b1711d1cfcedab2ef7\": container with ID starting with fde4c645dfb146fa707a4545d70e5bcf05ff688dee2e28b1711d1cfcedab2ef7 not found: ID does not exist" Nov 24 08:00:50 crc kubenswrapper[4809]: I1124 08:00:50.127681 4809 scope.go:117] "RemoveContainer" containerID="1ade393e694b53bd26850e9bed1d6dc98d9a2a57d846fbbe53cdaa1972b3e32c" Nov 24 08:00:50 crc kubenswrapper[4809]: E1124 08:00:50.129506 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ade393e694b53bd26850e9bed1d6dc98d9a2a57d846fbbe53cdaa1972b3e32c\": container with ID starting with 1ade393e694b53bd26850e9bed1d6dc98d9a2a57d846fbbe53cdaa1972b3e32c not found: ID does not exist" containerID="1ade393e694b53bd26850e9bed1d6dc98d9a2a57d846fbbe53cdaa1972b3e32c" Nov 24 08:00:50 crc kubenswrapper[4809]: I1124 08:00:50.129564 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ade393e694b53bd26850e9bed1d6dc98d9a2a57d846fbbe53cdaa1972b3e32c"} err="failed to get container status \"1ade393e694b53bd26850e9bed1d6dc98d9a2a57d846fbbe53cdaa1972b3e32c\": rpc error: code = NotFound desc = could not find container \"1ade393e694b53bd26850e9bed1d6dc98d9a2a57d846fbbe53cdaa1972b3e32c\": container with ID starting with 1ade393e694b53bd26850e9bed1d6dc98d9a2a57d846fbbe53cdaa1972b3e32c not found: ID does not exist" Nov 24 08:00:50 crc kubenswrapper[4809]: I1124 08:00:50.129596 4809 scope.go:117] "RemoveContainer" containerID="43b3c9ed4165a96f89aab8fdd98c98aaae3d8391dd5fe97e6ded3c4dbe3a6630" Nov 24 08:00:50 crc kubenswrapper[4809]: E1124 08:00:50.131142 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43b3c9ed4165a96f89aab8fdd98c98aaae3d8391dd5fe97e6ded3c4dbe3a6630\": container with ID starting with 43b3c9ed4165a96f89aab8fdd98c98aaae3d8391dd5fe97e6ded3c4dbe3a6630 not found: ID does not exist" containerID="43b3c9ed4165a96f89aab8fdd98c98aaae3d8391dd5fe97e6ded3c4dbe3a6630" Nov 24 08:00:50 crc kubenswrapper[4809]: I1124 08:00:50.131166 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43b3c9ed4165a96f89aab8fdd98c98aaae3d8391dd5fe97e6ded3c4dbe3a6630"} err="failed to get container status \"43b3c9ed4165a96f89aab8fdd98c98aaae3d8391dd5fe97e6ded3c4dbe3a6630\": rpc error: code = NotFound desc = could not find container \"43b3c9ed4165a96f89aab8fdd98c98aaae3d8391dd5fe97e6ded3c4dbe3a6630\": container with ID starting with 43b3c9ed4165a96f89aab8fdd98c98aaae3d8391dd5fe97e6ded3c4dbe3a6630 not found: ID does not exist" Nov 24 08:00:50 crc kubenswrapper[4809]: I1124 08:00:50.902095 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f303785-0349-4fe0-b300-19efa68d61c4" path="/var/lib/kubelet/pods/0f303785-0349-4fe0-b300-19efa68d61c4/volumes" Nov 24 08:01:00 crc kubenswrapper[4809]: I1124 08:01:00.144551 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29399521-vzxdj"] Nov 24 08:01:00 crc kubenswrapper[4809]: E1124 08:01:00.145581 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f303785-0349-4fe0-b300-19efa68d61c4" containerName="extract-utilities" Nov 24 08:01:00 crc kubenswrapper[4809]: I1124 08:01:00.145596 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f303785-0349-4fe0-b300-19efa68d61c4" containerName="extract-utilities" Nov 24 08:01:00 crc kubenswrapper[4809]: E1124 08:01:00.145608 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f303785-0349-4fe0-b300-19efa68d61c4" containerName="extract-content" Nov 24 08:01:00 crc kubenswrapper[4809]: I1124 08:01:00.145614 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f303785-0349-4fe0-b300-19efa68d61c4" containerName="extract-content" Nov 24 08:01:00 crc kubenswrapper[4809]: E1124 08:01:00.145623 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f303785-0349-4fe0-b300-19efa68d61c4" containerName="registry-server" Nov 24 08:01:00 crc kubenswrapper[4809]: I1124 08:01:00.145631 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f303785-0349-4fe0-b300-19efa68d61c4" containerName="registry-server" Nov 24 08:01:00 crc kubenswrapper[4809]: I1124 08:01:00.145828 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f303785-0349-4fe0-b300-19efa68d61c4" containerName="registry-server" Nov 24 08:01:00 crc kubenswrapper[4809]: I1124 08:01:00.146536 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399521-vzxdj" Nov 24 08:01:00 crc kubenswrapper[4809]: I1124 08:01:00.165434 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29399521-vzxdj"] Nov 24 08:01:00 crc kubenswrapper[4809]: I1124 08:01:00.309466 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c14625c7-1bc7-4940-81d6-a35be5d251ef-fernet-keys\") pod \"keystone-cron-29399521-vzxdj\" (UID: \"c14625c7-1bc7-4940-81d6-a35be5d251ef\") " pod="openstack/keystone-cron-29399521-vzxdj" Nov 24 08:01:00 crc kubenswrapper[4809]: I1124 08:01:00.309519 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c14625c7-1bc7-4940-81d6-a35be5d251ef-combined-ca-bundle\") pod \"keystone-cron-29399521-vzxdj\" (UID: \"c14625c7-1bc7-4940-81d6-a35be5d251ef\") " pod="openstack/keystone-cron-29399521-vzxdj" Nov 24 08:01:00 crc kubenswrapper[4809]: I1124 08:01:00.309567 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5drlz\" (UniqueName: \"kubernetes.io/projected/c14625c7-1bc7-4940-81d6-a35be5d251ef-kube-api-access-5drlz\") pod \"keystone-cron-29399521-vzxdj\" (UID: \"c14625c7-1bc7-4940-81d6-a35be5d251ef\") " pod="openstack/keystone-cron-29399521-vzxdj" Nov 24 08:01:00 crc kubenswrapper[4809]: I1124 08:01:00.309604 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c14625c7-1bc7-4940-81d6-a35be5d251ef-config-data\") pod \"keystone-cron-29399521-vzxdj\" (UID: \"c14625c7-1bc7-4940-81d6-a35be5d251ef\") " pod="openstack/keystone-cron-29399521-vzxdj" Nov 24 08:01:00 crc kubenswrapper[4809]: I1124 08:01:00.410962 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c14625c7-1bc7-4940-81d6-a35be5d251ef-fernet-keys\") pod \"keystone-cron-29399521-vzxdj\" (UID: \"c14625c7-1bc7-4940-81d6-a35be5d251ef\") " pod="openstack/keystone-cron-29399521-vzxdj" Nov 24 08:01:00 crc kubenswrapper[4809]: I1124 08:01:00.411308 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c14625c7-1bc7-4940-81d6-a35be5d251ef-combined-ca-bundle\") pod \"keystone-cron-29399521-vzxdj\" (UID: \"c14625c7-1bc7-4940-81d6-a35be5d251ef\") " pod="openstack/keystone-cron-29399521-vzxdj" Nov 24 08:01:00 crc kubenswrapper[4809]: I1124 08:01:00.411445 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5drlz\" (UniqueName: \"kubernetes.io/projected/c14625c7-1bc7-4940-81d6-a35be5d251ef-kube-api-access-5drlz\") pod \"keystone-cron-29399521-vzxdj\" (UID: \"c14625c7-1bc7-4940-81d6-a35be5d251ef\") " pod="openstack/keystone-cron-29399521-vzxdj" Nov 24 08:01:00 crc kubenswrapper[4809]: I1124 08:01:00.411551 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c14625c7-1bc7-4940-81d6-a35be5d251ef-config-data\") pod \"keystone-cron-29399521-vzxdj\" (UID: \"c14625c7-1bc7-4940-81d6-a35be5d251ef\") " pod="openstack/keystone-cron-29399521-vzxdj" Nov 24 08:01:00 crc kubenswrapper[4809]: I1124 08:01:00.417620 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c14625c7-1bc7-4940-81d6-a35be5d251ef-combined-ca-bundle\") pod \"keystone-cron-29399521-vzxdj\" (UID: \"c14625c7-1bc7-4940-81d6-a35be5d251ef\") " pod="openstack/keystone-cron-29399521-vzxdj" Nov 24 08:01:00 crc kubenswrapper[4809]: I1124 08:01:00.418096 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c14625c7-1bc7-4940-81d6-a35be5d251ef-config-data\") pod \"keystone-cron-29399521-vzxdj\" (UID: \"c14625c7-1bc7-4940-81d6-a35be5d251ef\") " pod="openstack/keystone-cron-29399521-vzxdj" Nov 24 08:01:00 crc kubenswrapper[4809]: I1124 08:01:00.418877 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c14625c7-1bc7-4940-81d6-a35be5d251ef-fernet-keys\") pod \"keystone-cron-29399521-vzxdj\" (UID: \"c14625c7-1bc7-4940-81d6-a35be5d251ef\") " pod="openstack/keystone-cron-29399521-vzxdj" Nov 24 08:01:00 crc kubenswrapper[4809]: I1124 08:01:00.429443 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5drlz\" (UniqueName: \"kubernetes.io/projected/c14625c7-1bc7-4940-81d6-a35be5d251ef-kube-api-access-5drlz\") pod \"keystone-cron-29399521-vzxdj\" (UID: \"c14625c7-1bc7-4940-81d6-a35be5d251ef\") " pod="openstack/keystone-cron-29399521-vzxdj" Nov 24 08:01:00 crc kubenswrapper[4809]: I1124 08:01:00.463407 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399521-vzxdj" Nov 24 08:01:00 crc kubenswrapper[4809]: I1124 08:01:00.920664 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29399521-vzxdj"] Nov 24 08:01:01 crc kubenswrapper[4809]: I1124 08:01:01.129672 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399521-vzxdj" event={"ID":"c14625c7-1bc7-4940-81d6-a35be5d251ef","Type":"ContainerStarted","Data":"40951f4f6f5f7a4f5506585a9ab648330a4396fe0409856ef49dd654b9adaf55"} Nov 24 08:01:01 crc kubenswrapper[4809]: I1124 08:01:01.188637 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-8fgt8_7265684f-a24a-40d3-8eb5-eb2da39cb870/nmstate-console-plugin/0.log" Nov 24 08:01:01 crc kubenswrapper[4809]: I1124 08:01:01.264913 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-5dl97_ef4d3282-be25-4395-899d-92b0fc8a0a40/nmstate-handler/0.log" Nov 24 08:01:01 crc kubenswrapper[4809]: I1124 08:01:01.405110 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-782jg_c00579c8-50fc-4b4e-8ddf-79ccdc0647ec/kube-rbac-proxy/0.log" Nov 24 08:01:01 crc kubenswrapper[4809]: I1124 08:01:01.420094 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-782jg_c00579c8-50fc-4b4e-8ddf-79ccdc0647ec/nmstate-metrics/0.log" Nov 24 08:01:01 crc kubenswrapper[4809]: I1124 08:01:01.565599 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-hbbwq_dec4efb0-1f05-416f-8e23-773aaab020df/nmstate-operator/0.log" Nov 24 08:01:01 crc kubenswrapper[4809]: I1124 08:01:01.644733 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-7fhk7_5c12932d-5d71-4a4f-9dc5-bbf20584f029/nmstate-webhook/0.log" Nov 24 08:01:02 crc kubenswrapper[4809]: I1124 08:01:02.150315 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399521-vzxdj" event={"ID":"c14625c7-1bc7-4940-81d6-a35be5d251ef","Type":"ContainerStarted","Data":"45e94c1f4c4a5c37198fef0e9f1b228c64623bdcac13924f1fa2ae7041d0e667"} Nov 24 08:01:02 crc kubenswrapper[4809]: I1124 08:01:02.175142 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29399521-vzxdj" podStartSLOduration=2.175124659 podStartE2EDuration="2.175124659s" podCreationTimestamp="2025-11-24 08:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:01:02.166078383 +0000 UTC m=+3962.066670098" watchObservedRunningTime="2025-11-24 08:01:02.175124659 +0000 UTC m=+3962.075716384" Nov 24 08:01:04 crc kubenswrapper[4809]: I1124 08:01:04.171232 4809 generic.go:334] "Generic (PLEG): container finished" podID="c14625c7-1bc7-4940-81d6-a35be5d251ef" containerID="45e94c1f4c4a5c37198fef0e9f1b228c64623bdcac13924f1fa2ae7041d0e667" exitCode=0 Nov 24 08:01:04 crc kubenswrapper[4809]: I1124 08:01:04.171355 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399521-vzxdj" event={"ID":"c14625c7-1bc7-4940-81d6-a35be5d251ef","Type":"ContainerDied","Data":"45e94c1f4c4a5c37198fef0e9f1b228c64623bdcac13924f1fa2ae7041d0e667"} Nov 24 08:01:05 crc kubenswrapper[4809]: I1124 08:01:05.538851 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399521-vzxdj" Nov 24 08:01:05 crc kubenswrapper[4809]: I1124 08:01:05.711179 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c14625c7-1bc7-4940-81d6-a35be5d251ef-config-data\") pod \"c14625c7-1bc7-4940-81d6-a35be5d251ef\" (UID: \"c14625c7-1bc7-4940-81d6-a35be5d251ef\") " Nov 24 08:01:05 crc kubenswrapper[4809]: I1124 08:01:05.711308 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c14625c7-1bc7-4940-81d6-a35be5d251ef-combined-ca-bundle\") pod \"c14625c7-1bc7-4940-81d6-a35be5d251ef\" (UID: \"c14625c7-1bc7-4940-81d6-a35be5d251ef\") " Nov 24 08:01:05 crc kubenswrapper[4809]: I1124 08:01:05.711695 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5drlz\" (UniqueName: \"kubernetes.io/projected/c14625c7-1bc7-4940-81d6-a35be5d251ef-kube-api-access-5drlz\") pod \"c14625c7-1bc7-4940-81d6-a35be5d251ef\" (UID: \"c14625c7-1bc7-4940-81d6-a35be5d251ef\") " Nov 24 08:01:05 crc kubenswrapper[4809]: I1124 08:01:05.711835 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c14625c7-1bc7-4940-81d6-a35be5d251ef-fernet-keys\") pod \"c14625c7-1bc7-4940-81d6-a35be5d251ef\" (UID: \"c14625c7-1bc7-4940-81d6-a35be5d251ef\") " Nov 24 08:01:05 crc kubenswrapper[4809]: I1124 08:01:05.717478 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c14625c7-1bc7-4940-81d6-a35be5d251ef-kube-api-access-5drlz" (OuterVolumeSpecName: "kube-api-access-5drlz") pod "c14625c7-1bc7-4940-81d6-a35be5d251ef" (UID: "c14625c7-1bc7-4940-81d6-a35be5d251ef"). InnerVolumeSpecName "kube-api-access-5drlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:01:05 crc kubenswrapper[4809]: I1124 08:01:05.722095 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c14625c7-1bc7-4940-81d6-a35be5d251ef-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c14625c7-1bc7-4940-81d6-a35be5d251ef" (UID: "c14625c7-1bc7-4940-81d6-a35be5d251ef"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:01:05 crc kubenswrapper[4809]: I1124 08:01:05.744116 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c14625c7-1bc7-4940-81d6-a35be5d251ef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c14625c7-1bc7-4940-81d6-a35be5d251ef" (UID: "c14625c7-1bc7-4940-81d6-a35be5d251ef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:01:05 crc kubenswrapper[4809]: I1124 08:01:05.774685 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c14625c7-1bc7-4940-81d6-a35be5d251ef-config-data" (OuterVolumeSpecName: "config-data") pod "c14625c7-1bc7-4940-81d6-a35be5d251ef" (UID: "c14625c7-1bc7-4940-81d6-a35be5d251ef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:01:05 crc kubenswrapper[4809]: I1124 08:01:05.813829 4809 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c14625c7-1bc7-4940-81d6-a35be5d251ef-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 08:01:05 crc kubenswrapper[4809]: I1124 08:01:05.813871 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c14625c7-1bc7-4940-81d6-a35be5d251ef-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:01:05 crc kubenswrapper[4809]: I1124 08:01:05.813883 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c14625c7-1bc7-4940-81d6-a35be5d251ef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:01:05 crc kubenswrapper[4809]: I1124 08:01:05.813898 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5drlz\" (UniqueName: \"kubernetes.io/projected/c14625c7-1bc7-4940-81d6-a35be5d251ef-kube-api-access-5drlz\") on node \"crc\" DevicePath \"\"" Nov 24 08:01:06 crc kubenswrapper[4809]: I1124 08:01:06.189356 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399521-vzxdj" event={"ID":"c14625c7-1bc7-4940-81d6-a35be5d251ef","Type":"ContainerDied","Data":"40951f4f6f5f7a4f5506585a9ab648330a4396fe0409856ef49dd654b9adaf55"} Nov 24 08:01:06 crc kubenswrapper[4809]: I1124 08:01:06.189697 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40951f4f6f5f7a4f5506585a9ab648330a4396fe0409856ef49dd654b9adaf55" Nov 24 08:01:06 crc kubenswrapper[4809]: I1124 08:01:06.189420 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399521-vzxdj" Nov 24 08:01:15 crc kubenswrapper[4809]: I1124 08:01:15.002003 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-8p4jp_00f3a099-57a2-4da1-bb22-12c48c359ab7/kube-rbac-proxy/0.log" Nov 24 08:01:15 crc kubenswrapper[4809]: I1124 08:01:15.137272 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-8p4jp_00f3a099-57a2-4da1-bb22-12c48c359ab7/controller/0.log" Nov 24 08:01:15 crc kubenswrapper[4809]: I1124 08:01:15.210800 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-b922n_ab5d28f0-1848-47f6-bdc8-444e258af818/frr-k8s-webhook-server/0.log" Nov 24 08:01:15 crc kubenswrapper[4809]: I1124 08:01:15.314254 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-frr-files/0.log" Nov 24 08:01:15 crc kubenswrapper[4809]: I1124 08:01:15.924480 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-frr-files/0.log" Nov 24 08:01:16 crc kubenswrapper[4809]: I1124 08:01:16.082315 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-metrics/0.log" Nov 24 08:01:16 crc kubenswrapper[4809]: I1124 08:01:16.090197 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-reloader/0.log" Nov 24 08:01:16 crc kubenswrapper[4809]: I1124 08:01:16.102570 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-reloader/0.log" Nov 24 08:01:16 crc kubenswrapper[4809]: I1124 08:01:16.216292 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-frr-files/0.log" Nov 24 08:01:16 crc kubenswrapper[4809]: I1124 08:01:16.276138 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-metrics/0.log" Nov 24 08:01:16 crc kubenswrapper[4809]: I1124 08:01:16.331192 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-metrics/0.log" Nov 24 08:01:16 crc kubenswrapper[4809]: I1124 08:01:16.333326 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-reloader/0.log" Nov 24 08:01:16 crc kubenswrapper[4809]: I1124 08:01:16.458166 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-reloader/0.log" Nov 24 08:01:16 crc kubenswrapper[4809]: I1124 08:01:16.466002 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-frr-files/0.log" Nov 24 08:01:16 crc kubenswrapper[4809]: I1124 08:01:16.513454 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/cp-metrics/0.log" Nov 24 08:01:16 crc kubenswrapper[4809]: I1124 08:01:16.522392 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/controller/0.log" Nov 24 08:01:16 crc kubenswrapper[4809]: I1124 08:01:16.640738 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/frr-metrics/0.log" Nov 24 08:01:16 crc kubenswrapper[4809]: I1124 08:01:16.734454 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/kube-rbac-proxy/0.log" Nov 24 08:01:16 crc kubenswrapper[4809]: I1124 08:01:16.737031 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/kube-rbac-proxy-frr/0.log" Nov 24 08:01:16 crc kubenswrapper[4809]: I1124 08:01:16.880728 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/reloader/0.log" Nov 24 08:01:16 crc kubenswrapper[4809]: I1124 08:01:16.942838 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7d8556447d-spn4w_a9a36b6c-e79c-4180-9118-146f3d280f15/manager/0.log" Nov 24 08:01:17 crc kubenswrapper[4809]: I1124 08:01:17.727222 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5969b4cdf8-zmvtt_baa671e1-bc87-45e6-8c04-33f408289fb4/webhook-server/0.log" Nov 24 08:01:17 crc kubenswrapper[4809]: I1124 08:01:17.956826 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-fpksh_13297f30-8d99-42d0-9760-9f88705b691f/kube-rbac-proxy/0.log" Nov 24 08:01:18 crc kubenswrapper[4809]: I1124 08:01:18.134150 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z5gsd_c263eb1f-d6cc-47d5-a4e8-6e5b12a7566b/frr/0.log" Nov 24 08:01:18 crc kubenswrapper[4809]: I1124 08:01:18.285458 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-fpksh_13297f30-8d99-42d0-9760-9f88705b691f/speaker/0.log" Nov 24 08:01:28 crc kubenswrapper[4809]: I1124 08:01:28.898905 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv_4bac6145-551f-4994-83ea-86181d6719ae/util/0.log" Nov 24 08:01:29 crc kubenswrapper[4809]: I1124 08:01:29.097565 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv_4bac6145-551f-4994-83ea-86181d6719ae/pull/0.log" Nov 24 08:01:29 crc kubenswrapper[4809]: I1124 08:01:29.098015 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv_4bac6145-551f-4994-83ea-86181d6719ae/pull/0.log" Nov 24 08:01:29 crc kubenswrapper[4809]: I1124 08:01:29.100124 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv_4bac6145-551f-4994-83ea-86181d6719ae/util/0.log" Nov 24 08:01:29 crc kubenswrapper[4809]: I1124 08:01:29.266672 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv_4bac6145-551f-4994-83ea-86181d6719ae/pull/0.log" Nov 24 08:01:29 crc kubenswrapper[4809]: I1124 08:01:29.290581 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv_4bac6145-551f-4994-83ea-86181d6719ae/util/0.log" Nov 24 08:01:29 crc kubenswrapper[4809]: I1124 08:01:29.301523 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ep9lrv_4bac6145-551f-4994-83ea-86181d6719ae/extract/0.log" Nov 24 08:01:29 crc kubenswrapper[4809]: I1124 08:01:29.455405 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7lv57_a56840cb-71cf-40df-8df8-c91866948a88/extract-utilities/0.log" Nov 24 08:01:29 crc kubenswrapper[4809]: I1124 08:01:29.615293 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7lv57_a56840cb-71cf-40df-8df8-c91866948a88/extract-content/0.log" Nov 24 08:01:29 crc kubenswrapper[4809]: I1124 08:01:29.621424 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7lv57_a56840cb-71cf-40df-8df8-c91866948a88/extract-utilities/0.log" Nov 24 08:01:29 crc kubenswrapper[4809]: I1124 08:01:29.671676 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7lv57_a56840cb-71cf-40df-8df8-c91866948a88/extract-content/0.log" Nov 24 08:01:29 crc kubenswrapper[4809]: I1124 08:01:29.871445 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7lv57_a56840cb-71cf-40df-8df8-c91866948a88/extract-content/0.log" Nov 24 08:01:29 crc kubenswrapper[4809]: I1124 08:01:29.912370 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7lv57_a56840cb-71cf-40df-8df8-c91866948a88/extract-utilities/0.log" Nov 24 08:01:30 crc kubenswrapper[4809]: I1124 08:01:30.114952 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ws6wx_422d766e-1186-4b50-b09c-2fcced744b5d/extract-utilities/0.log" Nov 24 08:01:30 crc kubenswrapper[4809]: I1124 08:01:30.280945 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ws6wx_422d766e-1186-4b50-b09c-2fcced744b5d/extract-utilities/0.log" Nov 24 08:01:30 crc kubenswrapper[4809]: I1124 08:01:30.333623 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7lv57_a56840cb-71cf-40df-8df8-c91866948a88/registry-server/0.log" Nov 24 08:01:30 crc kubenswrapper[4809]: I1124 08:01:30.341492 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ws6wx_422d766e-1186-4b50-b09c-2fcced744b5d/extract-content/0.log" Nov 24 08:01:30 crc kubenswrapper[4809]: I1124 08:01:30.351991 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ws6wx_422d766e-1186-4b50-b09c-2fcced744b5d/extract-content/0.log" Nov 24 08:01:30 crc kubenswrapper[4809]: I1124 08:01:30.511687 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ws6wx_422d766e-1186-4b50-b09c-2fcced744b5d/extract-utilities/0.log" Nov 24 08:01:30 crc kubenswrapper[4809]: I1124 08:01:30.556202 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ws6wx_422d766e-1186-4b50-b09c-2fcced744b5d/extract-content/0.log" Nov 24 08:01:30 crc kubenswrapper[4809]: I1124 08:01:30.706154 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8_7feb42ed-e415-44b8-a837-da74daf4aa52/util/0.log" Nov 24 08:01:30 crc kubenswrapper[4809]: I1124 08:01:30.819073 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ws6wx_422d766e-1186-4b50-b09c-2fcced744b5d/registry-server/0.log" Nov 24 08:01:30 crc kubenswrapper[4809]: I1124 08:01:30.910073 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8_7feb42ed-e415-44b8-a837-da74daf4aa52/util/0.log" Nov 24 08:01:30 crc kubenswrapper[4809]: I1124 08:01:30.920958 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8_7feb42ed-e415-44b8-a837-da74daf4aa52/pull/0.log" Nov 24 08:01:30 crc kubenswrapper[4809]: I1124 08:01:30.941095 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8_7feb42ed-e415-44b8-a837-da74daf4aa52/pull/0.log" Nov 24 08:01:31 crc kubenswrapper[4809]: I1124 08:01:31.085751 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8_7feb42ed-e415-44b8-a837-da74daf4aa52/util/0.log" Nov 24 08:01:31 crc kubenswrapper[4809]: I1124 08:01:31.101124 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8_7feb42ed-e415-44b8-a837-da74daf4aa52/extract/0.log" Nov 24 08:01:31 crc kubenswrapper[4809]: I1124 08:01:31.107340 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wqhl8_7feb42ed-e415-44b8-a837-da74daf4aa52/pull/0.log" Nov 24 08:01:31 crc kubenswrapper[4809]: I1124 08:01:31.265513 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-zqwx5_4b7632a5-d284-4a30-89dc-c301a7d13609/marketplace-operator/0.log" Nov 24 08:01:31 crc kubenswrapper[4809]: I1124 08:01:31.288127 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-l9v7r_9586cac6-9d7e-4c04-9400-b0760540e65c/extract-utilities/0.log" Nov 24 08:01:31 crc kubenswrapper[4809]: I1124 08:01:31.476985 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-l9v7r_9586cac6-9d7e-4c04-9400-b0760540e65c/extract-content/0.log" Nov 24 08:01:31 crc kubenswrapper[4809]: I1124 08:01:31.487830 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-l9v7r_9586cac6-9d7e-4c04-9400-b0760540e65c/extract-content/0.log" Nov 24 08:01:31 crc kubenswrapper[4809]: I1124 08:01:31.495899 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-l9v7r_9586cac6-9d7e-4c04-9400-b0760540e65c/extract-utilities/0.log" Nov 24 08:01:31 crc kubenswrapper[4809]: I1124 08:01:31.647689 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-l9v7r_9586cac6-9d7e-4c04-9400-b0760540e65c/extract-utilities/0.log" Nov 24 08:01:31 crc kubenswrapper[4809]: I1124 08:01:31.808331 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-l9v7r_9586cac6-9d7e-4c04-9400-b0760540e65c/registry-server/0.log" Nov 24 08:01:31 crc kubenswrapper[4809]: I1124 08:01:31.877243 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-l9v7r_9586cac6-9d7e-4c04-9400-b0760540e65c/extract-content/0.log" Nov 24 08:01:31 crc kubenswrapper[4809]: I1124 08:01:31.980889 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nzgcn_17720a8a-d721-431b-8bc4-1efd5ac1a634/extract-utilities/0.log" Nov 24 08:01:32 crc kubenswrapper[4809]: I1124 08:01:32.200010 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nzgcn_17720a8a-d721-431b-8bc4-1efd5ac1a634/extract-utilities/0.log" Nov 24 08:01:32 crc kubenswrapper[4809]: I1124 08:01:32.222643 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nzgcn_17720a8a-d721-431b-8bc4-1efd5ac1a634/extract-content/0.log" Nov 24 08:01:32 crc kubenswrapper[4809]: I1124 08:01:32.251232 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nzgcn_17720a8a-d721-431b-8bc4-1efd5ac1a634/extract-content/0.log" Nov 24 08:01:32 crc kubenswrapper[4809]: I1124 08:01:32.396726 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nzgcn_17720a8a-d721-431b-8bc4-1efd5ac1a634/extract-utilities/0.log" Nov 24 08:01:32 crc kubenswrapper[4809]: I1124 08:01:32.421267 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nzgcn_17720a8a-d721-431b-8bc4-1efd5ac1a634/extract-content/0.log" Nov 24 08:01:32 crc kubenswrapper[4809]: I1124 08:01:32.622804 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nzgcn_17720a8a-d721-431b-8bc4-1efd5ac1a634/registry-server/0.log" Nov 24 08:02:18 crc kubenswrapper[4809]: I1124 08:02:18.043150 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:02:18 crc kubenswrapper[4809]: I1124 08:02:18.044257 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:02:20 crc kubenswrapper[4809]: I1124 08:02:20.229871 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-n2g58"] Nov 24 08:02:20 crc kubenswrapper[4809]: E1124 08:02:20.230395 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c14625c7-1bc7-4940-81d6-a35be5d251ef" containerName="keystone-cron" Nov 24 08:02:20 crc kubenswrapper[4809]: I1124 08:02:20.230407 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c14625c7-1bc7-4940-81d6-a35be5d251ef" containerName="keystone-cron" Nov 24 08:02:20 crc kubenswrapper[4809]: I1124 08:02:20.230618 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c14625c7-1bc7-4940-81d6-a35be5d251ef" containerName="keystone-cron" Nov 24 08:02:20 crc kubenswrapper[4809]: I1124 08:02:20.275261 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n2g58"] Nov 24 08:02:20 crc kubenswrapper[4809]: I1124 08:02:20.275588 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n2g58" Nov 24 08:02:20 crc kubenswrapper[4809]: I1124 08:02:20.356326 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2908bb4b-cadf-4a5f-aed6-2358c8d1acf9-catalog-content\") pod \"certified-operators-n2g58\" (UID: \"2908bb4b-cadf-4a5f-aed6-2358c8d1acf9\") " pod="openshift-marketplace/certified-operators-n2g58" Nov 24 08:02:20 crc kubenswrapper[4809]: I1124 08:02:20.356769 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jbh4\" (UniqueName: \"kubernetes.io/projected/2908bb4b-cadf-4a5f-aed6-2358c8d1acf9-kube-api-access-9jbh4\") pod \"certified-operators-n2g58\" (UID: \"2908bb4b-cadf-4a5f-aed6-2358c8d1acf9\") " pod="openshift-marketplace/certified-operators-n2g58" Nov 24 08:02:20 crc kubenswrapper[4809]: I1124 08:02:20.356901 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2908bb4b-cadf-4a5f-aed6-2358c8d1acf9-utilities\") pod \"certified-operators-n2g58\" (UID: \"2908bb4b-cadf-4a5f-aed6-2358c8d1acf9\") " pod="openshift-marketplace/certified-operators-n2g58" Nov 24 08:02:20 crc kubenswrapper[4809]: I1124 08:02:20.459669 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2908bb4b-cadf-4a5f-aed6-2358c8d1acf9-utilities\") pod \"certified-operators-n2g58\" (UID: \"2908bb4b-cadf-4a5f-aed6-2358c8d1acf9\") " pod="openshift-marketplace/certified-operators-n2g58" Nov 24 08:02:20 crc kubenswrapper[4809]: I1124 08:02:20.460076 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2908bb4b-cadf-4a5f-aed6-2358c8d1acf9-catalog-content\") pod \"certified-operators-n2g58\" (UID: \"2908bb4b-cadf-4a5f-aed6-2358c8d1acf9\") " pod="openshift-marketplace/certified-operators-n2g58" Nov 24 08:02:20 crc kubenswrapper[4809]: I1124 08:02:20.460142 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jbh4\" (UniqueName: \"kubernetes.io/projected/2908bb4b-cadf-4a5f-aed6-2358c8d1acf9-kube-api-access-9jbh4\") pod \"certified-operators-n2g58\" (UID: \"2908bb4b-cadf-4a5f-aed6-2358c8d1acf9\") " pod="openshift-marketplace/certified-operators-n2g58" Nov 24 08:02:20 crc kubenswrapper[4809]: I1124 08:02:20.460499 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2908bb4b-cadf-4a5f-aed6-2358c8d1acf9-utilities\") pod \"certified-operators-n2g58\" (UID: \"2908bb4b-cadf-4a5f-aed6-2358c8d1acf9\") " pod="openshift-marketplace/certified-operators-n2g58" Nov 24 08:02:20 crc kubenswrapper[4809]: I1124 08:02:20.460596 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2908bb4b-cadf-4a5f-aed6-2358c8d1acf9-catalog-content\") pod \"certified-operators-n2g58\" (UID: \"2908bb4b-cadf-4a5f-aed6-2358c8d1acf9\") " pod="openshift-marketplace/certified-operators-n2g58" Nov 24 08:02:20 crc kubenswrapper[4809]: I1124 08:02:20.482298 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jbh4\" (UniqueName: \"kubernetes.io/projected/2908bb4b-cadf-4a5f-aed6-2358c8d1acf9-kube-api-access-9jbh4\") pod \"certified-operators-n2g58\" (UID: \"2908bb4b-cadf-4a5f-aed6-2358c8d1acf9\") " pod="openshift-marketplace/certified-operators-n2g58" Nov 24 08:02:20 crc kubenswrapper[4809]: I1124 08:02:20.599875 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n2g58" Nov 24 08:02:21 crc kubenswrapper[4809]: I1124 08:02:21.083512 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n2g58"] Nov 24 08:02:21 crc kubenswrapper[4809]: I1124 08:02:21.918195 4809 generic.go:334] "Generic (PLEG): container finished" podID="2908bb4b-cadf-4a5f-aed6-2358c8d1acf9" containerID="00c6d8d87a2b20a031c68fb401ac3c6f94f4d959a60fb0421c5be92f5a07871e" exitCode=0 Nov 24 08:02:21 crc kubenswrapper[4809]: I1124 08:02:21.918261 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2g58" event={"ID":"2908bb4b-cadf-4a5f-aed6-2358c8d1acf9","Type":"ContainerDied","Data":"00c6d8d87a2b20a031c68fb401ac3c6f94f4d959a60fb0421c5be92f5a07871e"} Nov 24 08:02:21 crc kubenswrapper[4809]: I1124 08:02:21.918511 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2g58" event={"ID":"2908bb4b-cadf-4a5f-aed6-2358c8d1acf9","Type":"ContainerStarted","Data":"b472706bdaaa90e2f112296fd67905a2a42c457de2275f7998703123f9828306"} Nov 24 08:02:22 crc kubenswrapper[4809]: I1124 08:02:22.928320 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2g58" event={"ID":"2908bb4b-cadf-4a5f-aed6-2358c8d1acf9","Type":"ContainerStarted","Data":"c7e460457c77a1bc42706373d47bbc61ce97ecddeba7387d41a7ead5bd1fa2f3"} Nov 24 08:02:23 crc kubenswrapper[4809]: I1124 08:02:23.939980 4809 generic.go:334] "Generic (PLEG): container finished" podID="2908bb4b-cadf-4a5f-aed6-2358c8d1acf9" containerID="c7e460457c77a1bc42706373d47bbc61ce97ecddeba7387d41a7ead5bd1fa2f3" exitCode=0 Nov 24 08:02:23 crc kubenswrapper[4809]: I1124 08:02:23.940024 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2g58" event={"ID":"2908bb4b-cadf-4a5f-aed6-2358c8d1acf9","Type":"ContainerDied","Data":"c7e460457c77a1bc42706373d47bbc61ce97ecddeba7387d41a7ead5bd1fa2f3"} Nov 24 08:02:25 crc kubenswrapper[4809]: I1124 08:02:25.979078 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2g58" event={"ID":"2908bb4b-cadf-4a5f-aed6-2358c8d1acf9","Type":"ContainerStarted","Data":"d007f9c6d1c136bf46efea23c6c4e266406a338372e31c1e401da8f6507ced61"} Nov 24 08:02:26 crc kubenswrapper[4809]: I1124 08:02:25.999732 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-n2g58" podStartSLOduration=3.586343785 podStartE2EDuration="5.999714534s" podCreationTimestamp="2025-11-24 08:02:20 +0000 UTC" firstStartedPulling="2025-11-24 08:02:21.920347502 +0000 UTC m=+4041.820939207" lastFinishedPulling="2025-11-24 08:02:24.333718251 +0000 UTC m=+4044.234309956" observedRunningTime="2025-11-24 08:02:25.997254487 +0000 UTC m=+4045.897846192" watchObservedRunningTime="2025-11-24 08:02:25.999714534 +0000 UTC m=+4045.900306239" Nov 24 08:02:30 crc kubenswrapper[4809]: I1124 08:02:30.600953 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-n2g58" Nov 24 08:02:30 crc kubenswrapper[4809]: I1124 08:02:30.603121 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-n2g58" Nov 24 08:02:30 crc kubenswrapper[4809]: I1124 08:02:30.669370 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-n2g58" Nov 24 08:02:31 crc kubenswrapper[4809]: I1124 08:02:31.067676 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-n2g58" Nov 24 08:02:31 crc kubenswrapper[4809]: I1124 08:02:31.911349 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n2g58"] Nov 24 08:02:33 crc kubenswrapper[4809]: I1124 08:02:33.040832 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-n2g58" podUID="2908bb4b-cadf-4a5f-aed6-2358c8d1acf9" containerName="registry-server" containerID="cri-o://d007f9c6d1c136bf46efea23c6c4e266406a338372e31c1e401da8f6507ced61" gracePeriod=2 Nov 24 08:02:33 crc kubenswrapper[4809]: I1124 08:02:33.463760 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n2g58" Nov 24 08:02:33 crc kubenswrapper[4809]: I1124 08:02:33.518429 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2908bb4b-cadf-4a5f-aed6-2358c8d1acf9-catalog-content\") pod \"2908bb4b-cadf-4a5f-aed6-2358c8d1acf9\" (UID: \"2908bb4b-cadf-4a5f-aed6-2358c8d1acf9\") " Nov 24 08:02:33 crc kubenswrapper[4809]: I1124 08:02:33.518571 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2908bb4b-cadf-4a5f-aed6-2358c8d1acf9-utilities\") pod \"2908bb4b-cadf-4a5f-aed6-2358c8d1acf9\" (UID: \"2908bb4b-cadf-4a5f-aed6-2358c8d1acf9\") " Nov 24 08:02:33 crc kubenswrapper[4809]: I1124 08:02:33.518694 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jbh4\" (UniqueName: \"kubernetes.io/projected/2908bb4b-cadf-4a5f-aed6-2358c8d1acf9-kube-api-access-9jbh4\") pod \"2908bb4b-cadf-4a5f-aed6-2358c8d1acf9\" (UID: \"2908bb4b-cadf-4a5f-aed6-2358c8d1acf9\") " Nov 24 08:02:33 crc kubenswrapper[4809]: I1124 08:02:33.523265 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2908bb4b-cadf-4a5f-aed6-2358c8d1acf9-utilities" (OuterVolumeSpecName: "utilities") pod "2908bb4b-cadf-4a5f-aed6-2358c8d1acf9" (UID: "2908bb4b-cadf-4a5f-aed6-2358c8d1acf9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:02:33 crc kubenswrapper[4809]: I1124 08:02:33.527867 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2908bb4b-cadf-4a5f-aed6-2358c8d1acf9-kube-api-access-9jbh4" (OuterVolumeSpecName: "kube-api-access-9jbh4") pod "2908bb4b-cadf-4a5f-aed6-2358c8d1acf9" (UID: "2908bb4b-cadf-4a5f-aed6-2358c8d1acf9"). InnerVolumeSpecName "kube-api-access-9jbh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:02:33 crc kubenswrapper[4809]: I1124 08:02:33.619794 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2908bb4b-cadf-4a5f-aed6-2358c8d1acf9-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:02:33 crc kubenswrapper[4809]: I1124 08:02:33.619825 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jbh4\" (UniqueName: \"kubernetes.io/projected/2908bb4b-cadf-4a5f-aed6-2358c8d1acf9-kube-api-access-9jbh4\") on node \"crc\" DevicePath \"\"" Nov 24 08:02:33 crc kubenswrapper[4809]: I1124 08:02:33.675417 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2908bb4b-cadf-4a5f-aed6-2358c8d1acf9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2908bb4b-cadf-4a5f-aed6-2358c8d1acf9" (UID: "2908bb4b-cadf-4a5f-aed6-2358c8d1acf9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:02:33 crc kubenswrapper[4809]: I1124 08:02:33.730421 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2908bb4b-cadf-4a5f-aed6-2358c8d1acf9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:02:34 crc kubenswrapper[4809]: I1124 08:02:34.049802 4809 generic.go:334] "Generic (PLEG): container finished" podID="2908bb4b-cadf-4a5f-aed6-2358c8d1acf9" containerID="d007f9c6d1c136bf46efea23c6c4e266406a338372e31c1e401da8f6507ced61" exitCode=0 Nov 24 08:02:34 crc kubenswrapper[4809]: I1124 08:02:34.049841 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2g58" event={"ID":"2908bb4b-cadf-4a5f-aed6-2358c8d1acf9","Type":"ContainerDied","Data":"d007f9c6d1c136bf46efea23c6c4e266406a338372e31c1e401da8f6507ced61"} Nov 24 08:02:34 crc kubenswrapper[4809]: I1124 08:02:34.049866 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2g58" event={"ID":"2908bb4b-cadf-4a5f-aed6-2358c8d1acf9","Type":"ContainerDied","Data":"b472706bdaaa90e2f112296fd67905a2a42c457de2275f7998703123f9828306"} Nov 24 08:02:34 crc kubenswrapper[4809]: I1124 08:02:34.049883 4809 scope.go:117] "RemoveContainer" containerID="d007f9c6d1c136bf46efea23c6c4e266406a338372e31c1e401da8f6507ced61" Nov 24 08:02:34 crc kubenswrapper[4809]: I1124 08:02:34.050028 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n2g58" Nov 24 08:02:34 crc kubenswrapper[4809]: I1124 08:02:34.088455 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n2g58"] Nov 24 08:02:34 crc kubenswrapper[4809]: I1124 08:02:34.089582 4809 scope.go:117] "RemoveContainer" containerID="c7e460457c77a1bc42706373d47bbc61ce97ecddeba7387d41a7ead5bd1fa2f3" Nov 24 08:02:34 crc kubenswrapper[4809]: I1124 08:02:34.097463 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-n2g58"] Nov 24 08:02:34 crc kubenswrapper[4809]: I1124 08:02:34.118406 4809 scope.go:117] "RemoveContainer" containerID="00c6d8d87a2b20a031c68fb401ac3c6f94f4d959a60fb0421c5be92f5a07871e" Nov 24 08:02:34 crc kubenswrapper[4809]: I1124 08:02:34.168345 4809 scope.go:117] "RemoveContainer" containerID="d007f9c6d1c136bf46efea23c6c4e266406a338372e31c1e401da8f6507ced61" Nov 24 08:02:34 crc kubenswrapper[4809]: E1124 08:02:34.168884 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d007f9c6d1c136bf46efea23c6c4e266406a338372e31c1e401da8f6507ced61\": container with ID starting with d007f9c6d1c136bf46efea23c6c4e266406a338372e31c1e401da8f6507ced61 not found: ID does not exist" containerID="d007f9c6d1c136bf46efea23c6c4e266406a338372e31c1e401da8f6507ced61" Nov 24 08:02:34 crc kubenswrapper[4809]: I1124 08:02:34.168924 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d007f9c6d1c136bf46efea23c6c4e266406a338372e31c1e401da8f6507ced61"} err="failed to get container status \"d007f9c6d1c136bf46efea23c6c4e266406a338372e31c1e401da8f6507ced61\": rpc error: code = NotFound desc = could not find container \"d007f9c6d1c136bf46efea23c6c4e266406a338372e31c1e401da8f6507ced61\": container with ID starting with d007f9c6d1c136bf46efea23c6c4e266406a338372e31c1e401da8f6507ced61 not found: ID does not exist" Nov 24 08:02:34 crc kubenswrapper[4809]: I1124 08:02:34.168949 4809 scope.go:117] "RemoveContainer" containerID="c7e460457c77a1bc42706373d47bbc61ce97ecddeba7387d41a7ead5bd1fa2f3" Nov 24 08:02:34 crc kubenswrapper[4809]: E1124 08:02:34.169386 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7e460457c77a1bc42706373d47bbc61ce97ecddeba7387d41a7ead5bd1fa2f3\": container with ID starting with c7e460457c77a1bc42706373d47bbc61ce97ecddeba7387d41a7ead5bd1fa2f3 not found: ID does not exist" containerID="c7e460457c77a1bc42706373d47bbc61ce97ecddeba7387d41a7ead5bd1fa2f3" Nov 24 08:02:34 crc kubenswrapper[4809]: I1124 08:02:34.169433 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7e460457c77a1bc42706373d47bbc61ce97ecddeba7387d41a7ead5bd1fa2f3"} err="failed to get container status \"c7e460457c77a1bc42706373d47bbc61ce97ecddeba7387d41a7ead5bd1fa2f3\": rpc error: code = NotFound desc = could not find container \"c7e460457c77a1bc42706373d47bbc61ce97ecddeba7387d41a7ead5bd1fa2f3\": container with ID starting with c7e460457c77a1bc42706373d47bbc61ce97ecddeba7387d41a7ead5bd1fa2f3 not found: ID does not exist" Nov 24 08:02:34 crc kubenswrapper[4809]: I1124 08:02:34.169462 4809 scope.go:117] "RemoveContainer" containerID="00c6d8d87a2b20a031c68fb401ac3c6f94f4d959a60fb0421c5be92f5a07871e" Nov 24 08:02:34 crc kubenswrapper[4809]: E1124 08:02:34.170082 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00c6d8d87a2b20a031c68fb401ac3c6f94f4d959a60fb0421c5be92f5a07871e\": container with ID starting with 00c6d8d87a2b20a031c68fb401ac3c6f94f4d959a60fb0421c5be92f5a07871e not found: ID does not exist" containerID="00c6d8d87a2b20a031c68fb401ac3c6f94f4d959a60fb0421c5be92f5a07871e" Nov 24 08:02:34 crc kubenswrapper[4809]: I1124 08:02:34.170114 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00c6d8d87a2b20a031c68fb401ac3c6f94f4d959a60fb0421c5be92f5a07871e"} err="failed to get container status \"00c6d8d87a2b20a031c68fb401ac3c6f94f4d959a60fb0421c5be92f5a07871e\": rpc error: code = NotFound desc = could not find container \"00c6d8d87a2b20a031c68fb401ac3c6f94f4d959a60fb0421c5be92f5a07871e\": container with ID starting with 00c6d8d87a2b20a031c68fb401ac3c6f94f4d959a60fb0421c5be92f5a07871e not found: ID does not exist" Nov 24 08:02:34 crc kubenswrapper[4809]: I1124 08:02:34.907489 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2908bb4b-cadf-4a5f-aed6-2358c8d1acf9" path="/var/lib/kubelet/pods/2908bb4b-cadf-4a5f-aed6-2358c8d1acf9/volumes" Nov 24 08:02:48 crc kubenswrapper[4809]: I1124 08:02:48.043072 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:02:48 crc kubenswrapper[4809]: I1124 08:02:48.043611 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:03:13 crc kubenswrapper[4809]: I1124 08:03:13.592457 4809 generic.go:334] "Generic (PLEG): container finished" podID="d7b9f5b4-482c-406f-9f60-b5bc186d5d85" containerID="9cb24a6d45b957a626a18f5830beb1316ff0e9e77d7f41433a6d62cef50b75d3" exitCode=0 Nov 24 08:03:13 crc kubenswrapper[4809]: I1124 08:03:13.592547 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8fk5f/must-gather-kfhs5" event={"ID":"d7b9f5b4-482c-406f-9f60-b5bc186d5d85","Type":"ContainerDied","Data":"9cb24a6d45b957a626a18f5830beb1316ff0e9e77d7f41433a6d62cef50b75d3"} Nov 24 08:03:13 crc kubenswrapper[4809]: I1124 08:03:13.593633 4809 scope.go:117] "RemoveContainer" containerID="9cb24a6d45b957a626a18f5830beb1316ff0e9e77d7f41433a6d62cef50b75d3" Nov 24 08:03:14 crc kubenswrapper[4809]: I1124 08:03:14.595061 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8fk5f_must-gather-kfhs5_d7b9f5b4-482c-406f-9f60-b5bc186d5d85/gather/0.log" Nov 24 08:03:18 crc kubenswrapper[4809]: I1124 08:03:18.042922 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:03:18 crc kubenswrapper[4809]: I1124 08:03:18.044491 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:03:18 crc kubenswrapper[4809]: I1124 08:03:18.044632 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" Nov 24 08:03:18 crc kubenswrapper[4809]: I1124 08:03:18.045528 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"28e197e5ea01aea1c7445a8b46230eb9b3683e3bff6bd97e100500700dabc0ed"} pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 08:03:18 crc kubenswrapper[4809]: I1124 08:03:18.045706 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" containerID="cri-o://28e197e5ea01aea1c7445a8b46230eb9b3683e3bff6bd97e100500700dabc0ed" gracePeriod=600 Nov 24 08:03:18 crc kubenswrapper[4809]: I1124 08:03:18.647165 4809 generic.go:334] "Generic (PLEG): container finished" podID="be8050d5-1aea-453e-a06e-87702b0e856b" containerID="28e197e5ea01aea1c7445a8b46230eb9b3683e3bff6bd97e100500700dabc0ed" exitCode=0 Nov 24 08:03:18 crc kubenswrapper[4809]: I1124 08:03:18.647260 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerDied","Data":"28e197e5ea01aea1c7445a8b46230eb9b3683e3bff6bd97e100500700dabc0ed"} Nov 24 08:03:18 crc kubenswrapper[4809]: I1124 08:03:18.647642 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" event={"ID":"be8050d5-1aea-453e-a06e-87702b0e856b","Type":"ContainerStarted","Data":"5c1fcdbf654133e2c460b4b2e6cffdd3d699b883e0e93622bcde38a94a12b732"} Nov 24 08:03:18 crc kubenswrapper[4809]: I1124 08:03:18.647720 4809 scope.go:117] "RemoveContainer" containerID="2e28c43fa5010eaa17eb9e4822aa6e557804c1889cb9e1f8e8b6f736d983edb0" Nov 24 08:03:24 crc kubenswrapper[4809]: I1124 08:03:24.226464 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8fk5f/must-gather-kfhs5"] Nov 24 08:03:24 crc kubenswrapper[4809]: I1124 08:03:24.227845 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-8fk5f/must-gather-kfhs5" podUID="d7b9f5b4-482c-406f-9f60-b5bc186d5d85" containerName="copy" containerID="cri-o://a91f3fb40b90f5af6695afbd3688541f40c0d8068e5e5d7ac455b4b8c50d9fed" gracePeriod=2 Nov 24 08:03:24 crc kubenswrapper[4809]: I1124 08:03:24.232687 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8fk5f/must-gather-kfhs5"] Nov 24 08:03:24 crc kubenswrapper[4809]: I1124 08:03:24.657687 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8fk5f_must-gather-kfhs5_d7b9f5b4-482c-406f-9f60-b5bc186d5d85/copy/0.log" Nov 24 08:03:24 crc kubenswrapper[4809]: I1124 08:03:24.658575 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8fk5f/must-gather-kfhs5" Nov 24 08:03:24 crc kubenswrapper[4809]: I1124 08:03:24.709020 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8fk5f_must-gather-kfhs5_d7b9f5b4-482c-406f-9f60-b5bc186d5d85/copy/0.log" Nov 24 08:03:24 crc kubenswrapper[4809]: I1124 08:03:24.709439 4809 generic.go:334] "Generic (PLEG): container finished" podID="d7b9f5b4-482c-406f-9f60-b5bc186d5d85" containerID="a91f3fb40b90f5af6695afbd3688541f40c0d8068e5e5d7ac455b4b8c50d9fed" exitCode=143 Nov 24 08:03:24 crc kubenswrapper[4809]: I1124 08:03:24.709474 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8fk5f/must-gather-kfhs5" Nov 24 08:03:24 crc kubenswrapper[4809]: I1124 08:03:24.709508 4809 scope.go:117] "RemoveContainer" containerID="a91f3fb40b90f5af6695afbd3688541f40c0d8068e5e5d7ac455b4b8c50d9fed" Nov 24 08:03:24 crc kubenswrapper[4809]: I1124 08:03:24.732488 4809 scope.go:117] "RemoveContainer" containerID="9cb24a6d45b957a626a18f5830beb1316ff0e9e77d7f41433a6d62cef50b75d3" Nov 24 08:03:24 crc kubenswrapper[4809]: I1124 08:03:24.803329 4809 scope.go:117] "RemoveContainer" containerID="a91f3fb40b90f5af6695afbd3688541f40c0d8068e5e5d7ac455b4b8c50d9fed" Nov 24 08:03:24 crc kubenswrapper[4809]: E1124 08:03:24.803765 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a91f3fb40b90f5af6695afbd3688541f40c0d8068e5e5d7ac455b4b8c50d9fed\": container with ID starting with a91f3fb40b90f5af6695afbd3688541f40c0d8068e5e5d7ac455b4b8c50d9fed not found: ID does not exist" containerID="a91f3fb40b90f5af6695afbd3688541f40c0d8068e5e5d7ac455b4b8c50d9fed" Nov 24 08:03:24 crc kubenswrapper[4809]: I1124 08:03:24.803824 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a91f3fb40b90f5af6695afbd3688541f40c0d8068e5e5d7ac455b4b8c50d9fed"} err="failed to get container status \"a91f3fb40b90f5af6695afbd3688541f40c0d8068e5e5d7ac455b4b8c50d9fed\": rpc error: code = NotFound desc = could not find container \"a91f3fb40b90f5af6695afbd3688541f40c0d8068e5e5d7ac455b4b8c50d9fed\": container with ID starting with a91f3fb40b90f5af6695afbd3688541f40c0d8068e5e5d7ac455b4b8c50d9fed not found: ID does not exist" Nov 24 08:03:24 crc kubenswrapper[4809]: I1124 08:03:24.803849 4809 scope.go:117] "RemoveContainer" containerID="9cb24a6d45b957a626a18f5830beb1316ff0e9e77d7f41433a6d62cef50b75d3" Nov 24 08:03:24 crc kubenswrapper[4809]: E1124 08:03:24.804231 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cb24a6d45b957a626a18f5830beb1316ff0e9e77d7f41433a6d62cef50b75d3\": container with ID starting with 9cb24a6d45b957a626a18f5830beb1316ff0e9e77d7f41433a6d62cef50b75d3 not found: ID does not exist" containerID="9cb24a6d45b957a626a18f5830beb1316ff0e9e77d7f41433a6d62cef50b75d3" Nov 24 08:03:24 crc kubenswrapper[4809]: I1124 08:03:24.804271 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cb24a6d45b957a626a18f5830beb1316ff0e9e77d7f41433a6d62cef50b75d3"} err="failed to get container status \"9cb24a6d45b957a626a18f5830beb1316ff0e9e77d7f41433a6d62cef50b75d3\": rpc error: code = NotFound desc = could not find container \"9cb24a6d45b957a626a18f5830beb1316ff0e9e77d7f41433a6d62cef50b75d3\": container with ID starting with 9cb24a6d45b957a626a18f5830beb1316ff0e9e77d7f41433a6d62cef50b75d3 not found: ID does not exist" Nov 24 08:03:24 crc kubenswrapper[4809]: I1124 08:03:24.850254 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d7b9f5b4-482c-406f-9f60-b5bc186d5d85-must-gather-output\") pod \"d7b9f5b4-482c-406f-9f60-b5bc186d5d85\" (UID: \"d7b9f5b4-482c-406f-9f60-b5bc186d5d85\") " Nov 24 08:03:24 crc kubenswrapper[4809]: I1124 08:03:24.850483 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjvvd\" (UniqueName: \"kubernetes.io/projected/d7b9f5b4-482c-406f-9f60-b5bc186d5d85-kube-api-access-tjvvd\") pod \"d7b9f5b4-482c-406f-9f60-b5bc186d5d85\" (UID: \"d7b9f5b4-482c-406f-9f60-b5bc186d5d85\") " Nov 24 08:03:24 crc kubenswrapper[4809]: I1124 08:03:24.856455 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7b9f5b4-482c-406f-9f60-b5bc186d5d85-kube-api-access-tjvvd" (OuterVolumeSpecName: "kube-api-access-tjvvd") pod "d7b9f5b4-482c-406f-9f60-b5bc186d5d85" (UID: "d7b9f5b4-482c-406f-9f60-b5bc186d5d85"). InnerVolumeSpecName "kube-api-access-tjvvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:03:24 crc kubenswrapper[4809]: I1124 08:03:24.953850 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjvvd\" (UniqueName: \"kubernetes.io/projected/d7b9f5b4-482c-406f-9f60-b5bc186d5d85-kube-api-access-tjvvd\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:24 crc kubenswrapper[4809]: I1124 08:03:24.984087 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7b9f5b4-482c-406f-9f60-b5bc186d5d85-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "d7b9f5b4-482c-406f-9f60-b5bc186d5d85" (UID: "d7b9f5b4-482c-406f-9f60-b5bc186d5d85"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:03:25 crc kubenswrapper[4809]: I1124 08:03:25.059455 4809 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d7b9f5b4-482c-406f-9f60-b5bc186d5d85-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:26 crc kubenswrapper[4809]: I1124 08:03:26.920874 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7b9f5b4-482c-406f-9f60-b5bc186d5d85" path="/var/lib/kubelet/pods/d7b9f5b4-482c-406f-9f60-b5bc186d5d85/volumes" Nov 24 08:04:43 crc kubenswrapper[4809]: I1124 08:04:43.638273 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7j7px"] Nov 24 08:04:43 crc kubenswrapper[4809]: E1124 08:04:43.639333 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2908bb4b-cadf-4a5f-aed6-2358c8d1acf9" containerName="registry-server" Nov 24 08:04:43 crc kubenswrapper[4809]: I1124 08:04:43.639350 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2908bb4b-cadf-4a5f-aed6-2358c8d1acf9" containerName="registry-server" Nov 24 08:04:43 crc kubenswrapper[4809]: E1124 08:04:43.639377 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2908bb4b-cadf-4a5f-aed6-2358c8d1acf9" containerName="extract-utilities" Nov 24 08:04:43 crc kubenswrapper[4809]: I1124 08:04:43.639386 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2908bb4b-cadf-4a5f-aed6-2358c8d1acf9" containerName="extract-utilities" Nov 24 08:04:43 crc kubenswrapper[4809]: E1124 08:04:43.639416 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7b9f5b4-482c-406f-9f60-b5bc186d5d85" containerName="gather" Nov 24 08:04:43 crc kubenswrapper[4809]: I1124 08:04:43.639424 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7b9f5b4-482c-406f-9f60-b5bc186d5d85" containerName="gather" Nov 24 08:04:43 crc kubenswrapper[4809]: E1124 08:04:43.639436 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7b9f5b4-482c-406f-9f60-b5bc186d5d85" containerName="copy" Nov 24 08:04:43 crc kubenswrapper[4809]: I1124 08:04:43.639445 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7b9f5b4-482c-406f-9f60-b5bc186d5d85" containerName="copy" Nov 24 08:04:43 crc kubenswrapper[4809]: E1124 08:04:43.639458 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2908bb4b-cadf-4a5f-aed6-2358c8d1acf9" containerName="extract-content" Nov 24 08:04:43 crc kubenswrapper[4809]: I1124 08:04:43.639465 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2908bb4b-cadf-4a5f-aed6-2358c8d1acf9" containerName="extract-content" Nov 24 08:04:43 crc kubenswrapper[4809]: I1124 08:04:43.639704 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7b9f5b4-482c-406f-9f60-b5bc186d5d85" containerName="gather" Nov 24 08:04:43 crc kubenswrapper[4809]: I1124 08:04:43.639721 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7b9f5b4-482c-406f-9f60-b5bc186d5d85" containerName="copy" Nov 24 08:04:43 crc kubenswrapper[4809]: I1124 08:04:43.639737 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2908bb4b-cadf-4a5f-aed6-2358c8d1acf9" containerName="registry-server" Nov 24 08:04:43 crc kubenswrapper[4809]: I1124 08:04:43.641476 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7j7px" Nov 24 08:04:43 crc kubenswrapper[4809]: I1124 08:04:43.663014 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7j7px"] Nov 24 08:04:43 crc kubenswrapper[4809]: I1124 08:04:43.837349 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqljl\" (UniqueName: \"kubernetes.io/projected/b2b71ace-f329-4f13-b403-f3d3c1b6b735-kube-api-access-nqljl\") pod \"redhat-marketplace-7j7px\" (UID: \"b2b71ace-f329-4f13-b403-f3d3c1b6b735\") " pod="openshift-marketplace/redhat-marketplace-7j7px" Nov 24 08:04:43 crc kubenswrapper[4809]: I1124 08:04:43.837423 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b71ace-f329-4f13-b403-f3d3c1b6b735-catalog-content\") pod \"redhat-marketplace-7j7px\" (UID: \"b2b71ace-f329-4f13-b403-f3d3c1b6b735\") " pod="openshift-marketplace/redhat-marketplace-7j7px" Nov 24 08:04:43 crc kubenswrapper[4809]: I1124 08:04:43.837542 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b71ace-f329-4f13-b403-f3d3c1b6b735-utilities\") pod \"redhat-marketplace-7j7px\" (UID: \"b2b71ace-f329-4f13-b403-f3d3c1b6b735\") " pod="openshift-marketplace/redhat-marketplace-7j7px" Nov 24 08:04:43 crc kubenswrapper[4809]: I1124 08:04:43.938692 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b71ace-f329-4f13-b403-f3d3c1b6b735-utilities\") pod \"redhat-marketplace-7j7px\" (UID: \"b2b71ace-f329-4f13-b403-f3d3c1b6b735\") " pod="openshift-marketplace/redhat-marketplace-7j7px" Nov 24 08:04:43 crc kubenswrapper[4809]: I1124 08:04:43.938805 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqljl\" (UniqueName: \"kubernetes.io/projected/b2b71ace-f329-4f13-b403-f3d3c1b6b735-kube-api-access-nqljl\") pod \"redhat-marketplace-7j7px\" (UID: \"b2b71ace-f329-4f13-b403-f3d3c1b6b735\") " pod="openshift-marketplace/redhat-marketplace-7j7px" Nov 24 08:04:43 crc kubenswrapper[4809]: I1124 08:04:43.938855 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b71ace-f329-4f13-b403-f3d3c1b6b735-catalog-content\") pod \"redhat-marketplace-7j7px\" (UID: \"b2b71ace-f329-4f13-b403-f3d3c1b6b735\") " pod="openshift-marketplace/redhat-marketplace-7j7px" Nov 24 08:04:43 crc kubenswrapper[4809]: I1124 08:04:43.939385 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b71ace-f329-4f13-b403-f3d3c1b6b735-catalog-content\") pod \"redhat-marketplace-7j7px\" (UID: \"b2b71ace-f329-4f13-b403-f3d3c1b6b735\") " pod="openshift-marketplace/redhat-marketplace-7j7px" Nov 24 08:04:43 crc kubenswrapper[4809]: I1124 08:04:43.940483 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b71ace-f329-4f13-b403-f3d3c1b6b735-utilities\") pod \"redhat-marketplace-7j7px\" (UID: \"b2b71ace-f329-4f13-b403-f3d3c1b6b735\") " pod="openshift-marketplace/redhat-marketplace-7j7px" Nov 24 08:04:43 crc kubenswrapper[4809]: I1124 08:04:43.967683 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqljl\" (UniqueName: \"kubernetes.io/projected/b2b71ace-f329-4f13-b403-f3d3c1b6b735-kube-api-access-nqljl\") pod \"redhat-marketplace-7j7px\" (UID: \"b2b71ace-f329-4f13-b403-f3d3c1b6b735\") " pod="openshift-marketplace/redhat-marketplace-7j7px" Nov 24 08:04:43 crc kubenswrapper[4809]: I1124 08:04:43.975667 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7j7px" Nov 24 08:04:44 crc kubenswrapper[4809]: I1124 08:04:44.402421 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7j7px"] Nov 24 08:04:44 crc kubenswrapper[4809]: I1124 08:04:44.452085 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7j7px" event={"ID":"b2b71ace-f329-4f13-b403-f3d3c1b6b735","Type":"ContainerStarted","Data":"14e1e07f0293b334740eb35db9f1be7ff170fe6239d116e61ab3bfa0f4a0ac25"} Nov 24 08:04:45 crc kubenswrapper[4809]: I1124 08:04:45.476567 4809 generic.go:334] "Generic (PLEG): container finished" podID="b2b71ace-f329-4f13-b403-f3d3c1b6b735" containerID="554050d98a5b6556cb67155a882bfb68da6eda3e43a22c13d4b582d0e100d828" exitCode=0 Nov 24 08:04:45 crc kubenswrapper[4809]: I1124 08:04:45.476717 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7j7px" event={"ID":"b2b71ace-f329-4f13-b403-f3d3c1b6b735","Type":"ContainerDied","Data":"554050d98a5b6556cb67155a882bfb68da6eda3e43a22c13d4b582d0e100d828"} Nov 24 08:04:48 crc kubenswrapper[4809]: I1124 08:04:48.507194 4809 generic.go:334] "Generic (PLEG): container finished" podID="b2b71ace-f329-4f13-b403-f3d3c1b6b735" containerID="ac4f3c9e41ba525549047e6a72efd2f75dd7c4373251b9b0df8c8ae11ecd8638" exitCode=0 Nov 24 08:04:48 crc kubenswrapper[4809]: I1124 08:04:48.507265 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7j7px" event={"ID":"b2b71ace-f329-4f13-b403-f3d3c1b6b735","Type":"ContainerDied","Data":"ac4f3c9e41ba525549047e6a72efd2f75dd7c4373251b9b0df8c8ae11ecd8638"} Nov 24 08:04:49 crc kubenswrapper[4809]: I1124 08:04:49.950067 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d5k7v"] Nov 24 08:04:50 crc kubenswrapper[4809]: I1124 08:04:49.967145 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d5k7v" Nov 24 08:04:50 crc kubenswrapper[4809]: I1124 08:04:49.981106 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d5k7v"] Nov 24 08:04:50 crc kubenswrapper[4809]: I1124 08:04:50.172159 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f134ff81-9c2a-41bf-8f29-52d0eb812211-catalog-content\") pod \"community-operators-d5k7v\" (UID: \"f134ff81-9c2a-41bf-8f29-52d0eb812211\") " pod="openshift-marketplace/community-operators-d5k7v" Nov 24 08:04:50 crc kubenswrapper[4809]: I1124 08:04:50.172375 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f134ff81-9c2a-41bf-8f29-52d0eb812211-utilities\") pod \"community-operators-d5k7v\" (UID: \"f134ff81-9c2a-41bf-8f29-52d0eb812211\") " pod="openshift-marketplace/community-operators-d5k7v" Nov 24 08:04:50 crc kubenswrapper[4809]: I1124 08:04:50.172464 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmvdb\" (UniqueName: \"kubernetes.io/projected/f134ff81-9c2a-41bf-8f29-52d0eb812211-kube-api-access-jmvdb\") pod \"community-operators-d5k7v\" (UID: \"f134ff81-9c2a-41bf-8f29-52d0eb812211\") " pod="openshift-marketplace/community-operators-d5k7v" Nov 24 08:04:50 crc kubenswrapper[4809]: I1124 08:04:50.275110 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f134ff81-9c2a-41bf-8f29-52d0eb812211-catalog-content\") pod \"community-operators-d5k7v\" (UID: \"f134ff81-9c2a-41bf-8f29-52d0eb812211\") " pod="openshift-marketplace/community-operators-d5k7v" Nov 24 08:04:50 crc kubenswrapper[4809]: I1124 08:04:50.275227 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f134ff81-9c2a-41bf-8f29-52d0eb812211-utilities\") pod \"community-operators-d5k7v\" (UID: \"f134ff81-9c2a-41bf-8f29-52d0eb812211\") " pod="openshift-marketplace/community-operators-d5k7v" Nov 24 08:04:50 crc kubenswrapper[4809]: I1124 08:04:50.275282 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmvdb\" (UniqueName: \"kubernetes.io/projected/f134ff81-9c2a-41bf-8f29-52d0eb812211-kube-api-access-jmvdb\") pod \"community-operators-d5k7v\" (UID: \"f134ff81-9c2a-41bf-8f29-52d0eb812211\") " pod="openshift-marketplace/community-operators-d5k7v" Nov 24 08:04:50 crc kubenswrapper[4809]: I1124 08:04:50.275889 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f134ff81-9c2a-41bf-8f29-52d0eb812211-catalog-content\") pod \"community-operators-d5k7v\" (UID: \"f134ff81-9c2a-41bf-8f29-52d0eb812211\") " pod="openshift-marketplace/community-operators-d5k7v" Nov 24 08:04:50 crc kubenswrapper[4809]: I1124 08:04:50.276310 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f134ff81-9c2a-41bf-8f29-52d0eb812211-utilities\") pod \"community-operators-d5k7v\" (UID: \"f134ff81-9c2a-41bf-8f29-52d0eb812211\") " pod="openshift-marketplace/community-operators-d5k7v" Nov 24 08:04:50 crc kubenswrapper[4809]: I1124 08:04:50.299034 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmvdb\" (UniqueName: \"kubernetes.io/projected/f134ff81-9c2a-41bf-8f29-52d0eb812211-kube-api-access-jmvdb\") pod \"community-operators-d5k7v\" (UID: \"f134ff81-9c2a-41bf-8f29-52d0eb812211\") " pod="openshift-marketplace/community-operators-d5k7v" Nov 24 08:04:50 crc kubenswrapper[4809]: I1124 08:04:50.306116 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d5k7v" Nov 24 08:04:51 crc kubenswrapper[4809]: I1124 08:04:51.565570 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7j7px" event={"ID":"b2b71ace-f329-4f13-b403-f3d3c1b6b735","Type":"ContainerStarted","Data":"abd7f7ba7928239004a4b4003ec01a66e8c221fc8802ee144eec05a2c9a3119c"} Nov 24 08:04:51 crc kubenswrapper[4809]: I1124 08:04:51.585717 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7j7px" podStartSLOduration=2.98036467 podStartE2EDuration="8.585697468s" podCreationTimestamp="2025-11-24 08:04:43 +0000 UTC" firstStartedPulling="2025-11-24 08:04:45.479616719 +0000 UTC m=+4185.380208424" lastFinishedPulling="2025-11-24 08:04:51.084949517 +0000 UTC m=+4190.985541222" observedRunningTime="2025-11-24 08:04:51.57959439 +0000 UTC m=+4191.480186095" watchObservedRunningTime="2025-11-24 08:04:51.585697468 +0000 UTC m=+4191.486289173" Nov 24 08:04:52 crc kubenswrapper[4809]: I1124 08:04:52.318800 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d5k7v"] Nov 24 08:04:52 crc kubenswrapper[4809]: I1124 08:04:52.576597 4809 generic.go:334] "Generic (PLEG): container finished" podID="f134ff81-9c2a-41bf-8f29-52d0eb812211" containerID="4e84b79ab1a6d5caeef650701d2ec8d6b4050290ec431b8d7b79e6626968ce24" exitCode=0 Nov 24 08:04:52 crc kubenswrapper[4809]: I1124 08:04:52.576671 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5k7v" event={"ID":"f134ff81-9c2a-41bf-8f29-52d0eb812211","Type":"ContainerDied","Data":"4e84b79ab1a6d5caeef650701d2ec8d6b4050290ec431b8d7b79e6626968ce24"} Nov 24 08:04:52 crc kubenswrapper[4809]: I1124 08:04:52.577072 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5k7v" event={"ID":"f134ff81-9c2a-41bf-8f29-52d0eb812211","Type":"ContainerStarted","Data":"ce8136410dff383c5c6521641a676cd81410fb3795f141df45fcd3e3b9dd10ea"} Nov 24 08:04:53 crc kubenswrapper[4809]: I1124 08:04:53.976574 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7j7px" Nov 24 08:04:53 crc kubenswrapper[4809]: I1124 08:04:53.976890 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7j7px" Nov 24 08:04:54 crc kubenswrapper[4809]: I1124 08:04:54.031303 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7j7px" Nov 24 08:04:54 crc kubenswrapper[4809]: I1124 08:04:54.598219 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5k7v" event={"ID":"f134ff81-9c2a-41bf-8f29-52d0eb812211","Type":"ContainerStarted","Data":"18041993801cc314e15006cf2073f59437361114ecfaeff76f805f2f165baf04"} Nov 24 08:04:55 crc kubenswrapper[4809]: I1124 08:04:55.611502 4809 generic.go:334] "Generic (PLEG): container finished" podID="f134ff81-9c2a-41bf-8f29-52d0eb812211" containerID="18041993801cc314e15006cf2073f59437361114ecfaeff76f805f2f165baf04" exitCode=0 Nov 24 08:04:55 crc kubenswrapper[4809]: I1124 08:04:55.611578 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5k7v" event={"ID":"f134ff81-9c2a-41bf-8f29-52d0eb812211","Type":"ContainerDied","Data":"18041993801cc314e15006cf2073f59437361114ecfaeff76f805f2f165baf04"} Nov 24 08:04:56 crc kubenswrapper[4809]: I1124 08:04:56.623006 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5k7v" event={"ID":"f134ff81-9c2a-41bf-8f29-52d0eb812211","Type":"ContainerStarted","Data":"494d06439d9c489c6a932c9ea772776a596eca0fe7c13d6eb28bec220bf7f69f"} Nov 24 08:04:56 crc kubenswrapper[4809]: I1124 08:04:56.647411 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d5k7v" podStartSLOduration=4.074812309 podStartE2EDuration="7.647393526s" podCreationTimestamp="2025-11-24 08:04:49 +0000 UTC" firstStartedPulling="2025-11-24 08:04:52.578026327 +0000 UTC m=+4192.478618052" lastFinishedPulling="2025-11-24 08:04:56.150607564 +0000 UTC m=+4196.051199269" observedRunningTime="2025-11-24 08:04:56.642290675 +0000 UTC m=+4196.542882380" watchObservedRunningTime="2025-11-24 08:04:56.647393526 +0000 UTC m=+4196.547985231" Nov 24 08:05:00 crc kubenswrapper[4809]: I1124 08:05:00.306447 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d5k7v" Nov 24 08:05:00 crc kubenswrapper[4809]: I1124 08:05:00.307166 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d5k7v" Nov 24 08:05:00 crc kubenswrapper[4809]: I1124 08:05:00.363437 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d5k7v" Nov 24 08:05:04 crc kubenswrapper[4809]: I1124 08:05:04.027906 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7j7px" Nov 24 08:05:04 crc kubenswrapper[4809]: I1124 08:05:04.094919 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7j7px"] Nov 24 08:05:04 crc kubenswrapper[4809]: I1124 08:05:04.691323 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7j7px" podUID="b2b71ace-f329-4f13-b403-f3d3c1b6b735" containerName="registry-server" containerID="cri-o://abd7f7ba7928239004a4b4003ec01a66e8c221fc8802ee144eec05a2c9a3119c" gracePeriod=2 Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.112308 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7j7px" Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.264028 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b71ace-f329-4f13-b403-f3d3c1b6b735-utilities\") pod \"b2b71ace-f329-4f13-b403-f3d3c1b6b735\" (UID: \"b2b71ace-f329-4f13-b403-f3d3c1b6b735\") " Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.264145 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqljl\" (UniqueName: \"kubernetes.io/projected/b2b71ace-f329-4f13-b403-f3d3c1b6b735-kube-api-access-nqljl\") pod \"b2b71ace-f329-4f13-b403-f3d3c1b6b735\" (UID: \"b2b71ace-f329-4f13-b403-f3d3c1b6b735\") " Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.264205 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b71ace-f329-4f13-b403-f3d3c1b6b735-catalog-content\") pod \"b2b71ace-f329-4f13-b403-f3d3c1b6b735\" (UID: \"b2b71ace-f329-4f13-b403-f3d3c1b6b735\") " Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.265107 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2b71ace-f329-4f13-b403-f3d3c1b6b735-utilities" (OuterVolumeSpecName: "utilities") pod "b2b71ace-f329-4f13-b403-f3d3c1b6b735" (UID: "b2b71ace-f329-4f13-b403-f3d3c1b6b735"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.272280 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2b71ace-f329-4f13-b403-f3d3c1b6b735-kube-api-access-nqljl" (OuterVolumeSpecName: "kube-api-access-nqljl") pod "b2b71ace-f329-4f13-b403-f3d3c1b6b735" (UID: "b2b71ace-f329-4f13-b403-f3d3c1b6b735"). InnerVolumeSpecName "kube-api-access-nqljl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.282482 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2b71ace-f329-4f13-b403-f3d3c1b6b735-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2b71ace-f329-4f13-b403-f3d3c1b6b735" (UID: "b2b71ace-f329-4f13-b403-f3d3c1b6b735"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.365805 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b71ace-f329-4f13-b403-f3d3c1b6b735-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.366237 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqljl\" (UniqueName: \"kubernetes.io/projected/b2b71ace-f329-4f13-b403-f3d3c1b6b735-kube-api-access-nqljl\") on node \"crc\" DevicePath \"\"" Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.366251 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b71ace-f329-4f13-b403-f3d3c1b6b735-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.702401 4809 generic.go:334] "Generic (PLEG): container finished" podID="b2b71ace-f329-4f13-b403-f3d3c1b6b735" containerID="abd7f7ba7928239004a4b4003ec01a66e8c221fc8802ee144eec05a2c9a3119c" exitCode=0 Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.702467 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7j7px" event={"ID":"b2b71ace-f329-4f13-b403-f3d3c1b6b735","Type":"ContainerDied","Data":"abd7f7ba7928239004a4b4003ec01a66e8c221fc8802ee144eec05a2c9a3119c"} Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.702482 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7j7px" Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.702509 4809 scope.go:117] "RemoveContainer" containerID="abd7f7ba7928239004a4b4003ec01a66e8c221fc8802ee144eec05a2c9a3119c" Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.702497 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7j7px" event={"ID":"b2b71ace-f329-4f13-b403-f3d3c1b6b735","Type":"ContainerDied","Data":"14e1e07f0293b334740eb35db9f1be7ff170fe6239d116e61ab3bfa0f4a0ac25"} Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.733678 4809 scope.go:117] "RemoveContainer" containerID="ac4f3c9e41ba525549047e6a72efd2f75dd7c4373251b9b0df8c8ae11ecd8638" Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.745130 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7j7px"] Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.755064 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7j7px"] Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.765509 4809 scope.go:117] "RemoveContainer" containerID="554050d98a5b6556cb67155a882bfb68da6eda3e43a22c13d4b582d0e100d828" Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.804182 4809 scope.go:117] "RemoveContainer" containerID="abd7f7ba7928239004a4b4003ec01a66e8c221fc8802ee144eec05a2c9a3119c" Nov 24 08:05:05 crc kubenswrapper[4809]: E1124 08:05:05.804556 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abd7f7ba7928239004a4b4003ec01a66e8c221fc8802ee144eec05a2c9a3119c\": container with ID starting with abd7f7ba7928239004a4b4003ec01a66e8c221fc8802ee144eec05a2c9a3119c not found: ID does not exist" containerID="abd7f7ba7928239004a4b4003ec01a66e8c221fc8802ee144eec05a2c9a3119c" Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.804595 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abd7f7ba7928239004a4b4003ec01a66e8c221fc8802ee144eec05a2c9a3119c"} err="failed to get container status \"abd7f7ba7928239004a4b4003ec01a66e8c221fc8802ee144eec05a2c9a3119c\": rpc error: code = NotFound desc = could not find container \"abd7f7ba7928239004a4b4003ec01a66e8c221fc8802ee144eec05a2c9a3119c\": container with ID starting with abd7f7ba7928239004a4b4003ec01a66e8c221fc8802ee144eec05a2c9a3119c not found: ID does not exist" Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.804621 4809 scope.go:117] "RemoveContainer" containerID="ac4f3c9e41ba525549047e6a72efd2f75dd7c4373251b9b0df8c8ae11ecd8638" Nov 24 08:05:05 crc kubenswrapper[4809]: E1124 08:05:05.804902 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac4f3c9e41ba525549047e6a72efd2f75dd7c4373251b9b0df8c8ae11ecd8638\": container with ID starting with ac4f3c9e41ba525549047e6a72efd2f75dd7c4373251b9b0df8c8ae11ecd8638 not found: ID does not exist" containerID="ac4f3c9e41ba525549047e6a72efd2f75dd7c4373251b9b0df8c8ae11ecd8638" Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.804930 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac4f3c9e41ba525549047e6a72efd2f75dd7c4373251b9b0df8c8ae11ecd8638"} err="failed to get container status \"ac4f3c9e41ba525549047e6a72efd2f75dd7c4373251b9b0df8c8ae11ecd8638\": rpc error: code = NotFound desc = could not find container \"ac4f3c9e41ba525549047e6a72efd2f75dd7c4373251b9b0df8c8ae11ecd8638\": container with ID starting with ac4f3c9e41ba525549047e6a72efd2f75dd7c4373251b9b0df8c8ae11ecd8638 not found: ID does not exist" Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.804950 4809 scope.go:117] "RemoveContainer" containerID="554050d98a5b6556cb67155a882bfb68da6eda3e43a22c13d4b582d0e100d828" Nov 24 08:05:05 crc kubenswrapper[4809]: E1124 08:05:05.805293 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"554050d98a5b6556cb67155a882bfb68da6eda3e43a22c13d4b582d0e100d828\": container with ID starting with 554050d98a5b6556cb67155a882bfb68da6eda3e43a22c13d4b582d0e100d828 not found: ID does not exist" containerID="554050d98a5b6556cb67155a882bfb68da6eda3e43a22c13d4b582d0e100d828" Nov 24 08:05:05 crc kubenswrapper[4809]: I1124 08:05:05.805328 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"554050d98a5b6556cb67155a882bfb68da6eda3e43a22c13d4b582d0e100d828"} err="failed to get container status \"554050d98a5b6556cb67155a882bfb68da6eda3e43a22c13d4b582d0e100d828\": rpc error: code = NotFound desc = could not find container \"554050d98a5b6556cb67155a882bfb68da6eda3e43a22c13d4b582d0e100d828\": container with ID starting with 554050d98a5b6556cb67155a882bfb68da6eda3e43a22c13d4b582d0e100d828 not found: ID does not exist" Nov 24 08:05:06 crc kubenswrapper[4809]: I1124 08:05:06.905699 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2b71ace-f329-4f13-b403-f3d3c1b6b735" path="/var/lib/kubelet/pods/b2b71ace-f329-4f13-b403-f3d3c1b6b735/volumes" Nov 24 08:05:10 crc kubenswrapper[4809]: I1124 08:05:10.354767 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d5k7v" Nov 24 08:05:10 crc kubenswrapper[4809]: I1124 08:05:10.403113 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d5k7v"] Nov 24 08:05:10 crc kubenswrapper[4809]: I1124 08:05:10.746307 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d5k7v" podUID="f134ff81-9c2a-41bf-8f29-52d0eb812211" containerName="registry-server" containerID="cri-o://494d06439d9c489c6a932c9ea772776a596eca0fe7c13d6eb28bec220bf7f69f" gracePeriod=2 Nov 24 08:05:10 crc kubenswrapper[4809]: E1124 08:05:10.850032 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf134ff81_9c2a_41bf_8f29_52d0eb812211.slice/crio-conmon-494d06439d9c489c6a932c9ea772776a596eca0fe7c13d6eb28bec220bf7f69f.scope\": RecentStats: unable to find data in memory cache]" Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.179550 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d5k7v" Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.273028 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f134ff81-9c2a-41bf-8f29-52d0eb812211-catalog-content\") pod \"f134ff81-9c2a-41bf-8f29-52d0eb812211\" (UID: \"f134ff81-9c2a-41bf-8f29-52d0eb812211\") " Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.273246 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmvdb\" (UniqueName: \"kubernetes.io/projected/f134ff81-9c2a-41bf-8f29-52d0eb812211-kube-api-access-jmvdb\") pod \"f134ff81-9c2a-41bf-8f29-52d0eb812211\" (UID: \"f134ff81-9c2a-41bf-8f29-52d0eb812211\") " Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.273359 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f134ff81-9c2a-41bf-8f29-52d0eb812211-utilities\") pod \"f134ff81-9c2a-41bf-8f29-52d0eb812211\" (UID: \"f134ff81-9c2a-41bf-8f29-52d0eb812211\") " Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.274753 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f134ff81-9c2a-41bf-8f29-52d0eb812211-utilities" (OuterVolumeSpecName: "utilities") pod "f134ff81-9c2a-41bf-8f29-52d0eb812211" (UID: "f134ff81-9c2a-41bf-8f29-52d0eb812211"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.279480 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f134ff81-9c2a-41bf-8f29-52d0eb812211-kube-api-access-jmvdb" (OuterVolumeSpecName: "kube-api-access-jmvdb") pod "f134ff81-9c2a-41bf-8f29-52d0eb812211" (UID: "f134ff81-9c2a-41bf-8f29-52d0eb812211"). InnerVolumeSpecName "kube-api-access-jmvdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.322584 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f134ff81-9c2a-41bf-8f29-52d0eb812211-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f134ff81-9c2a-41bf-8f29-52d0eb812211" (UID: "f134ff81-9c2a-41bf-8f29-52d0eb812211"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.376444 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmvdb\" (UniqueName: \"kubernetes.io/projected/f134ff81-9c2a-41bf-8f29-52d0eb812211-kube-api-access-jmvdb\") on node \"crc\" DevicePath \"\"" Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.376493 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f134ff81-9c2a-41bf-8f29-52d0eb812211-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.376507 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f134ff81-9c2a-41bf-8f29-52d0eb812211-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.757230 4809 generic.go:334] "Generic (PLEG): container finished" podID="f134ff81-9c2a-41bf-8f29-52d0eb812211" containerID="494d06439d9c489c6a932c9ea772776a596eca0fe7c13d6eb28bec220bf7f69f" exitCode=0 Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.757282 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5k7v" event={"ID":"f134ff81-9c2a-41bf-8f29-52d0eb812211","Type":"ContainerDied","Data":"494d06439d9c489c6a932c9ea772776a596eca0fe7c13d6eb28bec220bf7f69f"} Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.757303 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d5k7v" Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.757319 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5k7v" event={"ID":"f134ff81-9c2a-41bf-8f29-52d0eb812211","Type":"ContainerDied","Data":"ce8136410dff383c5c6521641a676cd81410fb3795f141df45fcd3e3b9dd10ea"} Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.757345 4809 scope.go:117] "RemoveContainer" containerID="494d06439d9c489c6a932c9ea772776a596eca0fe7c13d6eb28bec220bf7f69f" Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.779907 4809 scope.go:117] "RemoveContainer" containerID="18041993801cc314e15006cf2073f59437361114ecfaeff76f805f2f165baf04" Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.788662 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d5k7v"] Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.795868 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-d5k7v"] Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.811868 4809 scope.go:117] "RemoveContainer" containerID="4e84b79ab1a6d5caeef650701d2ec8d6b4050290ec431b8d7b79e6626968ce24" Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.841063 4809 scope.go:117] "RemoveContainer" containerID="494d06439d9c489c6a932c9ea772776a596eca0fe7c13d6eb28bec220bf7f69f" Nov 24 08:05:11 crc kubenswrapper[4809]: E1124 08:05:11.841817 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"494d06439d9c489c6a932c9ea772776a596eca0fe7c13d6eb28bec220bf7f69f\": container with ID starting with 494d06439d9c489c6a932c9ea772776a596eca0fe7c13d6eb28bec220bf7f69f not found: ID does not exist" containerID="494d06439d9c489c6a932c9ea772776a596eca0fe7c13d6eb28bec220bf7f69f" Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.841872 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"494d06439d9c489c6a932c9ea772776a596eca0fe7c13d6eb28bec220bf7f69f"} err="failed to get container status \"494d06439d9c489c6a932c9ea772776a596eca0fe7c13d6eb28bec220bf7f69f\": rpc error: code = NotFound desc = could not find container \"494d06439d9c489c6a932c9ea772776a596eca0fe7c13d6eb28bec220bf7f69f\": container with ID starting with 494d06439d9c489c6a932c9ea772776a596eca0fe7c13d6eb28bec220bf7f69f not found: ID does not exist" Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.841905 4809 scope.go:117] "RemoveContainer" containerID="18041993801cc314e15006cf2073f59437361114ecfaeff76f805f2f165baf04" Nov 24 08:05:11 crc kubenswrapper[4809]: E1124 08:05:11.842305 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18041993801cc314e15006cf2073f59437361114ecfaeff76f805f2f165baf04\": container with ID starting with 18041993801cc314e15006cf2073f59437361114ecfaeff76f805f2f165baf04 not found: ID does not exist" containerID="18041993801cc314e15006cf2073f59437361114ecfaeff76f805f2f165baf04" Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.842351 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18041993801cc314e15006cf2073f59437361114ecfaeff76f805f2f165baf04"} err="failed to get container status \"18041993801cc314e15006cf2073f59437361114ecfaeff76f805f2f165baf04\": rpc error: code = NotFound desc = could not find container \"18041993801cc314e15006cf2073f59437361114ecfaeff76f805f2f165baf04\": container with ID starting with 18041993801cc314e15006cf2073f59437361114ecfaeff76f805f2f165baf04 not found: ID does not exist" Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.842376 4809 scope.go:117] "RemoveContainer" containerID="4e84b79ab1a6d5caeef650701d2ec8d6b4050290ec431b8d7b79e6626968ce24" Nov 24 08:05:11 crc kubenswrapper[4809]: E1124 08:05:11.842716 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e84b79ab1a6d5caeef650701d2ec8d6b4050290ec431b8d7b79e6626968ce24\": container with ID starting with 4e84b79ab1a6d5caeef650701d2ec8d6b4050290ec431b8d7b79e6626968ce24 not found: ID does not exist" containerID="4e84b79ab1a6d5caeef650701d2ec8d6b4050290ec431b8d7b79e6626968ce24" Nov 24 08:05:11 crc kubenswrapper[4809]: I1124 08:05:11.842784 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e84b79ab1a6d5caeef650701d2ec8d6b4050290ec431b8d7b79e6626968ce24"} err="failed to get container status \"4e84b79ab1a6d5caeef650701d2ec8d6b4050290ec431b8d7b79e6626968ce24\": rpc error: code = NotFound desc = could not find container \"4e84b79ab1a6d5caeef650701d2ec8d6b4050290ec431b8d7b79e6626968ce24\": container with ID starting with 4e84b79ab1a6d5caeef650701d2ec8d6b4050290ec431b8d7b79e6626968ce24 not found: ID does not exist" Nov 24 08:05:12 crc kubenswrapper[4809]: I1124 08:05:12.900865 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f134ff81-9c2a-41bf-8f29-52d0eb812211" path="/var/lib/kubelet/pods/f134ff81-9c2a-41bf-8f29-52d0eb812211/volumes" Nov 24 08:05:18 crc kubenswrapper[4809]: I1124 08:05:18.043529 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:05:18 crc kubenswrapper[4809]: I1124 08:05:18.044089 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:05:48 crc kubenswrapper[4809]: I1124 08:05:48.043608 4809 patch_prober.go:28] interesting pod/machine-config-daemon-dr8hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:05:48 crc kubenswrapper[4809]: I1124 08:05:48.044205 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dr8hv" podUID="be8050d5-1aea-453e-a06e-87702b0e856b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111010773024441 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111010773017356 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111000174016473 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111000174015443 5ustar corecore